Duplicate Content Errors
-
Ok, old fat client developer new at SEO so I apologize if this is obvious.
I have 4 errors in one of my campaigns. two are duplicate content and two are duplicate title.
Here is the duplicate title error
Rare Currency And Old Paper Money Values and Information.
http://www.antiquebanknotes.com/Rare Currency And Old Paper Money Values and Information.
http://www.antiquebanknotes.com/Default.aspxSo, my question is... What do I need to do to make this right? They are the same page. in my page load for default.aspx I have this:
this.Title = "Rare Currency And Old Paper Money Values and Information.";
And it occurs only once...
-
They offer II7 but I didn't see any point to it...
Now it's dawning on me I may need to reconsider.
-
I dont think it works on IIS6
There is your problem.
You can fix your www problem by code, see tutorials, but fixing the default page is not so easy. as in ASP.Net you can not detect the difference between domain.com and domain.com/default.aspx
Best thing you can do is make sure all your internal links point to domain.com and not domain.com/default.aspx any external links pointing to default.aspx will be watsed, but sicne no internal links point to default.aspx you are unlikely to get any mopre external links pointing to it.
GoDaddy much be using old servers.
-
It's II6 hosted on Godaddy. I will see if I can get my IIS restarted.
-
Cant see anything wrong with web.config
I should of asked before if you are using IIS7?
if so try to debug with the article http://blogs.iis.net/ruslany/archive/2008/10/30/debug-and-troubleshoot-rewrite-rules-easily.aspx
I have the same code as you in many sites all working fine
you may need to restart IIS?
-
nor is the non www redirected to the www
the code i gave you should be inside the configuration tag
<configuration></configuration>
place here
and you should only have one <system.webserver>tag</system.webserver>
if you want you cany send me your web.config and ill have a look
-
Ok, I have that in place... if I am testing right, the default.aspx isn't working. Maybe it's a hosting issue?
-
What this line is saying is, if not antiquebanknotes.com then redirect, note the negate=true
<add input="{HTTP_HOST}" pattern="^.antiquebanknotes.com$" negate="true"></add>
What you should have is
<system.webserver><rewrite><rules><rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^www.antiquebanknotes.com$" negate="true"></add></conditions>
<action type="Redirect" url="http://www.antiquebanknotes.com/{R:1}"></action></match></rule>
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule></rules></rewrite></system.webserver>Try that, let me know when you have it in place and I will test it for you.
-
I thought this pattern:
was ok for me bcause I am looking to take AntiqueBanknotes.com and make it into www.AntiqueBankNotes etc
I think my hostname redirect is working ok.... it's the default.aspx redir that isn't working. I have it in the same Rules section.
Look ok?
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule> -
You have an error above, you should have
pattern="^www.antiquebanknotes.com$"
not pattern="^antiquebanknotes\.com$"
The other rule does work i just tested it
When adding it place stright under the last rule inside the same rules tag, there should only be one rules tag ```
-
Sorry, I forgot to add that I put that rule in before but it doesn't seem to be working for me. Not sure why but sadly it's time for my day job so I will take a look tonight.
-
Add the rule on this page to it
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-issues-involving-the-default-pageit should work no worries
-
Ah, I see... I have added the snippets to the web config for default page and www vs non www pages. it seems like the www issue is working correctly. I will have to work on the default issue as it seems to ignore my rule.
This is what I have done on that score.
<rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^antiquebanknotes.com$" negate="true"></add></conditions>
<action type="Redirect" url="<a href=" http:="" www.antiquebanknotes.com="" {r:1"="">http://www.AntiqueBanknotes.com/{R:1}" />
</action></match></rule>Thanks for all your help!
-
Using iis writes to teh web.config file fior you, the tutorials show you the code it writes, you can just copy it into the web.config
There is also some tuorials on how to make a httpmodule that works much like the global.asax file but intercepts the requerst before it reaches the website, much better preformace.
Also i cheked your non www and it resolved to this
http://www.antiquebanknotes.com/antiquebanknotes/default.aspx
I notive above you dont have "" on the first url, is that just a typo?
this tutorial uses the httpmodule, it resolves the other way from www to nonwww, but you can alter it
http://perthseocompany.com.au/seo/tutorials/using-ihttpmodule-c-sharp
or you can just edit the web.config as I sugested above
-
Hi Alan,I appreciate the links and now I know what the problem is... sadly though I am not able to access IIS as this is running on Godaddy... I guess I will have to implement something in the global.asax file?I have this code in the global.asax to handle the www and non www page issue. string request = HttpContext.Current.Request.Url.ToString().ToLower();if (request.Contains("http://antiquebanknotes.com")){ HttpContext.Current.Response.Clear(); HttpContext.Current.Response.Status = "301 Moved Permanently"; HttpContext.Current.Response.AddHeader("Location", request.Replace (http://antiquebanknotes.com", "http://www.antiquebanknotes.com));}
-
They are seen as 2 different pages to search engines adn spliting your rank,
Here is a tutoprial just for this problem
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-issues-involving-the-default-page
You may want to have a look at this one also for yourt www and non www pages.
http://perthseocompany.com.au/seo/tutorials/how-to-fix-canonical-domain-name-issues
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate Content
I am trying to get a handle on how to fix and control a large amount of duplicate content I keep getting on my Moz Reports. The main area where this comes up is for duplicate page content and duplicate title tags ... thousands of them. I partially understand the source of the problem. My site mixes free content with content that requires a login. I think if I were to change my crawl settings to eliminate the login and index the paid content it would lower the quantity of duplicate pages and help me identify the true duplicate pages because a large number of duplicates occur at the site login. Unfortunately, it's not simple in my case because last year I encountered a problem when migrating my archives into a new CMS. The app in the CMS that migrated the data caused a large amount of data truncation Which means that I am piecing together my archives of approximately 5,000 articles. It also means that much of the piecing together process requires me to keep the former app that manages the articles to find where certain articles were truncated and to copy the text that followed the truncation and complete the articles. So far, I have restored about half of the archives which is time-consuming tedious work. My question is if anyone knows a more efficient way of identifying and editing duplicate pages and title tags?
Technical SEO | | Prop650 -
Handling of Duplicate Content
I just recently signed and joined the moz.com system. During the initial report for our web site it shows we have lots of duplicate content. The web site is real estate based and we are loading IDX listings from other brokerages into our site. If though these listings look alike, they are not. Each has their own photos, description and addresses. So why are they appear as duplicates – I would assume that they are all too closely related. Lots for Sale primarily – and it looks like lazy agents have 4 or 5 lots and input the description the same. Unfortunately for us, part of the IDX agreement is that you cannot pick and choose which listings to load and you cannot change the content. You are either all in or you cannot use the system. How should one manage duplicate content like this? Or should we ignore it? Out of 1500+ listings on our web site it shows 40 of them are duplicates.
Technical SEO | | TIM_DOTCOM0 -
Why is Coyscape showing content duplication error even after implementing 301 redirect ?
We are maintaining the corporate website of one of our prestigious clients "FineTech Toolings" (http://www.finetechtoolings.in). Recently I had raised a question regarding "2 websites running paralley in 2 diferent domains, i.e. 1 organisation having 2 different websites on 2 different domains". Recently my domain changed from http://www.finetechtoolings.co.in to http://www.finetechtoolings.in via 301 redirect, but still I am facing content duplication issue as per Copyscape. Hence I am having a small doubt regarding the same. Please note the following question very carefully and provide me the exact problem and the solution for the same: Even though I have implemented 301 redirect (http://www.finetechtoolings.co.in is redirected to http://www.finetechtoolings.in), which is completely ok as per the SEO rules, why is copyscape still showing that duplicate content exists in the former website? I think I am clear enough with my question.
Technical SEO | | KDKini0 -
Duplicate Content Issue WWW and Non WWW
One of my sites got hit with duplicate content a while ago because Google seemed to be considering hhtp, https, www, and non ww versions of the site all different sites. We thought we fixed it, but for some reason https://www and just https:// are giving us duplicate content again. I can't seem to figure out why it keeps doing this. The url is https://bandsonabudget.com if any of you want to see if you can figure out why I am still having this issue.
Technical SEO | | Michael4g1 -
Quickview popup duplicate content
Hi We have an eccomerce site. We just added to the product list view a quickview tab - when you roll mouse over it a popup window with the product image and short description shows up - is this a problem of duplicate content( its the same content that's on the product pages except there we also have a long detailed description) - t is done with javascript. Thanks!
Technical SEO | | henya0 -
Would Google Call These Pages Duplicate Content?
Our Web store, http://www.audiobooksonline.com/index.html, has struggled with duplicate content issues for some time. One aspect of duplicate content is a page like this: http://www.audiobooksonline.com/out-of-publication-audio-books-book-audiobook-audiobooks.html. When an audio book title goes out-of-publication we keep the page at our store and display a http://www.audiobooksonline.com/out-of-publication-audio-books-book-audiobook-audiobooks.html whenever a visitor attempts to visit a specific title that is OOP. There are several thousand OOP pages. Would Google consider these OOP pages duplicate content?
Technical SEO | | lbohen0 -
Rel=canonical overkill on duplicate content?
Our site has many different health centers - many of which contain duplicate content since there is topic crossover between health centers. I am using rel canonical to deal with this. My question is this: Is there a tipping point for duplicate content where Google might begin to penalize a site even if it has the rel canonical tags in place on cloned content? As an extreme example, a site could have 10 pieces of original content, but could then clone and organize this content in 5 different directories across the site each with a new url. This would ultimately result in the site having more "cloned" content than original content. Is this at all problematic even if the rel canonical is in place on all cloned content? Thanks in advance for any replies. Eric
Technical SEO | | Eric_Lifescript0 -
Duplicate Content - That Old Chestnut!!!
Hi Guys, Hope all is well, I have a question if I may? I have several articles which we have written and I want to try and find out the best way to post these but I have a number of concerns. I am hoping to use the content in an attempt to increase the Kudos of our site by providing quality content and hopefully receiving decent back links. 1. In terms of duplicate content should I only post it on one place or should I post the article in several places? Also where would you say the top 5 or 10 places would be? These are articles on XML, Social Media & Back Links. 2. Can I post the article on another blog or article directory and post it on my websites blog or is this a bad idea? A million thanks for any guidance. Kind Regards, C
Technical SEO | | fenwaymedia0