Duplicate Content Errors
-
Ok, old fat client developer new at SEO so I apologize if this is obvious.
I have 4 errors in one of my campaigns. two are duplicate content and two are duplicate title.
Here is the duplicate title error
Rare Currency And Old Paper Money Values and Information.
http://www.antiquebanknotes.com/Rare Currency And Old Paper Money Values and Information.
http://www.antiquebanknotes.com/Default.aspxSo, my question is... What do I need to do to make this right? They are the same page. in my page load for default.aspx I have this:
this.Title = "Rare Currency And Old Paper Money Values and Information.";
And it occurs only once...
-
They offer II7 but I didn't see any point to it...
Now it's dawning on me I may need to reconsider.
-
I dont think it works on IIS6
There is your problem.
You can fix your www problem by code, see tutorials, but fixing the default page is not so easy. as in ASP.Net you can not detect the difference between domain.com and domain.com/default.aspx
Best thing you can do is make sure all your internal links point to domain.com and not domain.com/default.aspx any external links pointing to default.aspx will be watsed, but sicne no internal links point to default.aspx you are unlikely to get any mopre external links pointing to it.
GoDaddy much be using old servers.
-
It's II6 hosted on Godaddy. I will see if I can get my IIS restarted.
-
Cant see anything wrong with web.config
I should of asked before if you are using IIS7?
if so try to debug with the article http://blogs.iis.net/ruslany/archive/2008/10/30/debug-and-troubleshoot-rewrite-rules-easily.aspx
I have the same code as you in many sites all working fine
you may need to restart IIS?
-
nor is the non www redirected to the www
the code i gave you should be inside the configuration tag
<configuration></configuration>
place here
and you should only have one <system.webserver>tag</system.webserver>
if you want you cany send me your web.config and ill have a look
-
Ok, I have that in place... if I am testing right, the default.aspx isn't working. Maybe it's a hosting issue?
-
What this line is saying is, if not antiquebanknotes.com then redirect, note the negate=true
<add input="{HTTP_HOST}" pattern="^.antiquebanknotes.com$" negate="true"></add>
What you should have is
<system.webserver><rewrite><rules><rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^www.antiquebanknotes.com$" negate="true"></add></conditions>
<action type="Redirect" url="http://www.antiquebanknotes.com/{R:1}"></action></match></rule>
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule></rules></rewrite></system.webserver>Try that, let me know when you have it in place and I will test it for you.
-
I thought this pattern:
was ok for me bcause I am looking to take AntiqueBanknotes.com and make it into www.AntiqueBankNotes etc
I think my hostname redirect is working ok.... it's the default.aspx redir that isn't working. I have it in the same Rules section.
Look ok?
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule> -
You have an error above, you should have
pattern="^www.antiquebanknotes.com$"
not pattern="^antiquebanknotes\.com$"
The other rule does work i just tested it
When adding it place stright under the last rule inside the same rules tag, there should only be one rules tag ```
-
Sorry, I forgot to add that I put that rule in before but it doesn't seem to be working for me. Not sure why but sadly it's time for my day job so I will take a look tonight.
-
Add the rule on this page to it
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-issues-involving-the-default-pageit should work no worries
-
Ah, I see... I have added the snippets to the web config for default page and www vs non www pages. it seems like the www issue is working correctly. I will have to work on the default issue as it seems to ignore my rule.
This is what I have done on that score.
<rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^antiquebanknotes.com$" negate="true"></add></conditions>
<action type="Redirect" url="<a href=" http:="" www.antiquebanknotes.com="" {r:1"="">http://www.AntiqueBanknotes.com/{R:1}" />
</action></match></rule>Thanks for all your help!
-
Using iis writes to teh web.config file fior you, the tutorials show you the code it writes, you can just copy it into the web.config
There is also some tuorials on how to make a httpmodule that works much like the global.asax file but intercepts the requerst before it reaches the website, much better preformace.
Also i cheked your non www and it resolved to this
http://www.antiquebanknotes.com/antiquebanknotes/default.aspx
I notive above you dont have "" on the first url, is that just a typo?
this tutorial uses the httpmodule, it resolves the other way from www to nonwww, but you can alter it
http://perthseocompany.com.au/seo/tutorials/using-ihttpmodule-c-sharp
or you can just edit the web.config as I sugested above
-
Hi Alan,I appreciate the links and now I know what the problem is... sadly though I am not able to access IIS as this is running on Godaddy... I guess I will have to implement something in the global.asax file?I have this code in the global.asax to handle the www and non www page issue. string request = HttpContext.Current.Request.Url.ToString().ToLower();if (request.Contains("http://antiquebanknotes.com")){ HttpContext.Current.Response.Clear(); HttpContext.Current.Response.Status = "301 Moved Permanently"; HttpContext.Current.Response.AddHeader("Location", request.Replace (http://antiquebanknotes.com", "http://www.antiquebanknotes.com));}
-
They are seen as 2 different pages to search engines adn spliting your rank,
Here is a tutoprial just for this problem
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-issues-involving-the-default-page
You may want to have a look at this one also for yourt www and non www pages.
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-domain-name-issues
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Who gets punished for duplicate content?
What happens if two domains have duplicate content? Do both domains get punished for it, or just one? If so, which one?
Technical SEO | | Tobii-Dynavox0 -
Duplicate content on report
Hi, I just had my Moz Campaign scan 10K pages out of which 2K were duplicate content and URL's are http://www.Somesite.com/modal/register?destination=question%2F37201 http://www.Somesite.com/modal/register?destination=question%2F37490 And the title for all 2K is "Register" How can i deal with this as all my pages have the register link and login and when done it comes back to the same page where we left and that it actually not duplicate but we need to deal with it propely thanks
Technical SEO | | mtthompsons0 -
Duplicate content in product listing
We have "duplicate content" warning in our moz report which mostly revolve around our product listing (eCommerce site) where various filters return 0 results (and hence show the same content on the page). Do you think those need to be addressed, and if so how would you prevent product listing filters that appearing as duplicate content pages? should we use rel=canonical or actually change the content on the page?
Technical SEO | | erangalp0 -
Image centric site and duplicate content issues
We have a site that has very little text, the main purpose of the site is to allow users to find inspiration through images. 1000s of images come to us each week to be processed by our editorial team, so as part of our process we select a subset of the best images and process those with titles, alt text, tags, etc. We still host the other images and users can find them through galleries that link to the process and unprocessed image pages. Due to the lack of information on the unprocessed images, we are having lots of duplicate content issues (The layout of all the image pages are the same, and there isn't any unique text to differentiate the pages. The only changing factor is the image itself in each page) Any suggestions on how to resolve this issue, will be greatly appreciated.
Technical SEO | | wedlinkmedia0 -
Duplicate page/Title content - Where?
Hi, I have just run a crawl on a new clients site, and there is several 'duplicate page content' and 'Duplicate Page Title'' issues. But I cannot find any duplicate content. And to make matters worse. The actual report has confused me. Just for example the about us page is showing in both reports and for both under 'Other URLs' it is showing 1? Why? Does this mean there is 1 other page with duplicate page title? or duplicate page content? Where are the pages that have the duplicate page titles, or duplicate page content? I have run scans using other software and a copyscape scan. And apart from missing page titles, I cannot find any page that has duplicate titles or content. I can find % percentages of pages with similar/same page titles/content. But this is only partial and contextually correct. So I understand that SEO Moz may pick percentage of content, which is fine, and therefore note that there is duplicate content/page titles. But I cannot seem to figure out where I would the source of the duplicate content/page titles. As there is only 1 listed in both reports for 'Other URLs' Hopefully my long question, has not confused. many thanks in advance for any help
Technical SEO | | wood1e20 -
Multiple Sites Duplicate Content Best Practice
Hi there, I have one client (atlantawidgets.com) who has a main site. But also has duplicate sites with different urls targeting specific geo areas. I.e. (widgetmakersinmarietta.com) Would it be best to go ahead and create a static home page at these add'l sites and make the rest of the site be nonindexed? Or should I go in and allow more pages to be indexed and change the content? If so how many, 3, 5, 8? I don't have tons of time at this point. 3)If I change content within the duplicate sites, what % do I need to change. Does switching the order of the sentences of the content count? Or does it need to be 100%fresh? Thanks everyone.
Technical SEO | | greenhornet770 -
Are RSS Feeds deemed duplicate content?
If a website content management system includes built-in feeds of different categories that the client can choose from, does that endanger them of having duplicate content if their categories are the same as another client's feed? These feeds appear on templated home page designs by default. Just trying to figure out how big of an issue these feeds are in terms of duplicate content across clients' sites. Should I be concerned? Obviously, there's other content on the home page besides the feed and have not really seen negative effects, but could it be impacting results?
Technical SEO | | KyleNeuberger0 -
Duplicate Content issue
I have been asked to review an old website to an identify opportunities for increasing search engine traffic. Whilst reviewing the site I came across a strange loop. On each page there is a link to printer friendly version: http://www.websitename.co.uk/index.php?pageid=7&printfriendly=yes That page also has a link to a printer friendly version http://www.websitename.co.uk/index.php?pageid=7&printfriendly=yes&printfriendly=yes and so on and so on....... Some of these pages are being included in Google's index. I appreciate that this can't be a good thing, however, I am not 100% sure as to the extent to which it is a bad thing and the priority that should be given to getting it sorted. Just wandering what views people have on the issues this may cause?
Technical SEO | | CPLDistribution0