Duplicate Content Errors
-
Ok, old fat client developer new at SEO so I apologize if this is obvious.
I have 4 errors in one of my campaigns. two are duplicate content and two are duplicate title.
Here is the duplicate title error
Rare Currency And Old Paper Money Values and Information.
http://www.antiquebanknotes.com/Rare Currency And Old Paper Money Values and Information.
http://www.antiquebanknotes.com/Default.aspxSo, my question is... What do I need to do to make this right? They are the same page. in my page load for default.aspx I have this:
this.Title = "Rare Currency And Old Paper Money Values and Information.";
And it occurs only once...
-
They offer II7 but I didn't see any point to it...
Now it's dawning on me I may need to reconsider.
-
I dont think it works on IIS6
There is your problem.
You can fix your www problem by code, see tutorials, but fixing the default page is not so easy. as in ASP.Net you can not detect the difference between domain.com and domain.com/default.aspx
Best thing you can do is make sure all your internal links point to domain.com and not domain.com/default.aspx any external links pointing to default.aspx will be watsed, but sicne no internal links point to default.aspx you are unlikely to get any mopre external links pointing to it.
GoDaddy much be using old servers.
-
It's II6 hosted on Godaddy. I will see if I can get my IIS restarted.
-
Cant see anything wrong with web.config
I should of asked before if you are using IIS7?
if so try to debug with the article http://blogs.iis.net/ruslany/archive/2008/10/30/debug-and-troubleshoot-rewrite-rules-easily.aspx
I have the same code as you in many sites all working fine
you may need to restart IIS?
-
nor is the non www redirected to the www
the code i gave you should be inside the configuration tag
<configuration></configuration>
place here
and you should only have one <system.webserver>tag</system.webserver>
if you want you cany send me your web.config and ill have a look
-
Ok, I have that in place... if I am testing right, the default.aspx isn't working. Maybe it's a hosting issue?
-
What this line is saying is, if not antiquebanknotes.com then redirect, note the negate=true
<add input="{HTTP_HOST}" pattern="^.antiquebanknotes.com$" negate="true"></add>
What you should have is
<system.webserver><rewrite><rules><rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^www.antiquebanknotes.com$" negate="true"></add></conditions>
<action type="Redirect" url="http://www.antiquebanknotes.com/{R:1}"></action></match></rule>
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule></rules></rewrite></system.webserver>Try that, let me know when you have it in place and I will test it for you.
-
I thought this pattern:
was ok for me bcause I am looking to take AntiqueBanknotes.com and make it into www.AntiqueBankNotes etc
I think my hostname redirect is working ok.... it's the default.aspx redir that isn't working. I have it in the same Rules section.
Look ok?
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule> -
You have an error above, you should have
pattern="^www.antiquebanknotes.com$"
not pattern="^antiquebanknotes\.com$"
The other rule does work i just tested it
When adding it place stright under the last rule inside the same rules tag, there should only be one rules tag ```
-
Sorry, I forgot to add that I put that rule in before but it doesn't seem to be working for me. Not sure why but sadly it's time for my day job so I will take a look tonight.
-
Add the rule on this page to it
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-issues-involving-the-default-pageit should work no worries
-
Ah, I see... I have added the snippets to the web config for default page and www vs non www pages. it seems like the www issue is working correctly. I will have to work on the default issue as it seems to ignore my rule.
This is what I have done on that score.
<rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^antiquebanknotes.com$" negate="true"></add></conditions>
<action type="Redirect" url="<a href=" http:="" www.antiquebanknotes.com="" {r:1"="">http://www.AntiqueBanknotes.com/{R:1}" />
</action></match></rule>Thanks for all your help!
-
Using iis writes to teh web.config file fior you, the tutorials show you the code it writes, you can just copy it into the web.config
There is also some tuorials on how to make a httpmodule that works much like the global.asax file but intercepts the requerst before it reaches the website, much better preformace.
Also i cheked your non www and it resolved to this
http://www.antiquebanknotes.com/antiquebanknotes/default.aspx
I notive above you dont have "" on the first url, is that just a typo?
this tutorial uses the httpmodule, it resolves the other way from www to nonwww, but you can alter it
http://perthseocompany.com.au/seo/tutorials/using-ihttpmodule-c-sharp
or you can just edit the web.config as I sugested above
-
Hi Alan,I appreciate the links and now I know what the problem is... sadly though I am not able to access IIS as this is running on Godaddy... I guess I will have to implement something in the global.asax file?I have this code in the global.asax to handle the www and non www page issue. string request = HttpContext.Current.Request.Url.ToString().ToLower();if (request.Contains("http://antiquebanknotes.com")){ HttpContext.Current.Response.Clear(); HttpContext.Current.Response.Status = "301 Moved Permanently"; HttpContext.Current.Response.AddHeader("Location", request.Replace (http://antiquebanknotes.com", "http://www.antiquebanknotes.com));}
-
They are seen as 2 different pages to search engines adn spliting your rank,
Here is a tutoprial just for this problem
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-issues-involving-the-default-page
You may want to have a look at this one also for yourt www and non www pages.
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-domain-name-issues
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate Content - Reverse Phone Directory
Hi, Until a few months ago, my client's site had about 600 pages. He decided to implement what is essentially a reverse phone directory/lookup tool. There are now about 10,000 reverse directory/lookup pages (.html), all with short and duplicate content except for the phone number and the caller name. Needless to say, I'm getting thousands of duplicate content errors. Are there tricks of the trade to deal with this? In nosing around, I've discovered that the pages are showing up in Google search results (when searching for a specific phone number), usually in the first or second position. Ideally, each page would have unique content, but that's next to impossible with 10,000 pages. One potential solution I've come up with is incorporating user-generated content into each page (maybe via Disqus?), which over time would make each page unique. I've also thought about suggesting that he move those pages onto a different domain. I'd appreciate any advice/suggestions, as well as any insights into the long-term repercussions of having so many dupes on the ranking of the 600 solidly unique pages on the site. Thanks in advance for your help!
Technical SEO | | sally580 -
Advice on Duplicate Page Content
We have many pages on our website and they all have the same template (we use a CMS) and at the code level, they are 90% the same. But the page content, title, meta description, and image used are different for all of them. For example - http://www.jumpstart.com/common/find-easter-eggs
Technical SEO | | jsmoz
http://www.jumpstart.com/common/recognize-the-rs We have many such pages. Does Google look at them all as duplicate page content? If yes, how do we deal with this?0 -
Localized domains and duplicate content
Hey guys, In my company we are launching a new website and there's an issue it's been bothering me for a while. I'm sure you guys can help me out. I already have a website, let's say ABC.com I'm preparing a localized version of that website for the uk so we'll launch ABC.co.uk Basically the websites are going to be exactly the same with the difference of the homepage. They have a slightly different proposition. Using GeoIP I will redirect the UK traffic to ABC.co.uk and the rest of the traffic will still visit .com website. May google penalize this? The site itself it will be almost the same but the homepage. This may count as duplicate content even if I'm geo-targeting different regions so they will never overlap. Thanks in advance for you advice
Technical SEO | | fabrizzio0 -
Issue: Duplicate Page Content
Hi All, I am getting warnings about duplicate page content. The pages are normally 'tag' pages. I have some blog posts tagged with multiple 'tags'. Does it really affect my site?. I am using wordpress and Yoast SEO plugin. Thanks
Technical SEO | | KLLC0 -
Errors - 7300 - Duplicate Page Content..Help me..
Hi, I just received the crawl report with 7300 errors of duplicate page content. Site built using php. list of errors will be like this.. http://xxxxx.com/channels/ http://xxxxx.com/channels/?page=1 http://xxxxxx.com/channels/?page=2 I am not good in coding and using readymade script for this website. could anyone guide me to fix this issue? Thanks.
Technical SEO | | vilambara0 -
Duplicate Page Content and Titles
A few weeks ago my error count went up for Duplicate Page Content and Titles. 4 errors in all. A week later the errors were gone... But now they are back. I made changes to the Webconfig over a month ago but nothing since. SEOmoz is telling me the duplicate content is this http://www.antiquebanknotes.com/ and http://www.antiquebanknotes.com Thanks for any advise! This is the relevant web.config. <rewrite><rules><rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^www.antiquebanknotes.com$" negate="true"></add></conditions>
Technical SEO | | Banknotes
<action type="Redirect" url="<a href=" http:="" www.antiquebanknotes.com="" {r:1"="">http://www.antiquebanknotes.com/{R:1}" />
</action></match></rule>
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule></rules></rewrite>0 -
How do I fix duplicate content with the home page?
This is probably SEO 101, but I'm unsure what to do here... Last week my weekly crawl diagnostics were off the chart because http:// was not resolving to http://www...fixed that but now it's saying I have duplicate content on: http://www.......com http://www.......com/index.php How do I fix this? Thanks in advance!
Technical SEO | | jgower0 -
Noindex duplicate content penalty?
We know that google now gives a penalty to a whole duplicate if it finds content it doesn't like or is duplicate content, but has anyone experienced a penalty from having duplicate content on their site which they have added noindex to? Would google still apply the penalty to the overall quality of the site even though they have been told to basically ignore the duplicate bit. Reason for asking is that I am looking to add a forum to one of my websites and no one likes a new forum. I have a script which can populate it with thousands of questions and answers pulled direct from Yahoo Answers. Obviously the forum wil be 100% duplicate content but I do not want it to rank for anyway anyway so if I noindex the forum pages hopefully it will not damage the rest of the site. In time, as the forum grows, all the duplicate posts will be deleted but it's hard to get people to use an empty forum so need to 'trick' them into thinking the section is very busy.
Technical SEO | | Grumpy_Carl0