About duplicate content
-
Hi i'm a new guy around here, but i'm having this problem in my website.
Using de Seomoz tools i ran a camping to my website, in results i get to many errors for duplicate conten, for example,
http://www.mysite/blue/index.html, so my question is,
what is the best way to resolve this problem, use a 301 or use the rel canonical tag?
Wich url will be consider for main url,
Thanks for yor help.
-
Hi,
I get duplicate content notifications on the following URLs:
and
www.mydomain.nl**/**
The / cause almost all my pages to show up as duplicate content. How to fix these?
Thnxs for the help!
-
I don't think some of the responses in this thread have given you adequate information to solve your problem. 301's and rel canonical are there to solve two very different problems, and when used correctly, can solve a lot of different SEO problems.
In your example you have two URLs which I am going to assume have the exact same information on them. Classic duplicate content situation. Ideally, I think you would want to delete one of these pages and create a 301 to redirect any users and links to the other page. This will focus all your content and links onto a single page and your PR and rankings will rise. I would choose to keep the page that has the better keywords in the URL, and no, it doesn't matter if you have the .html at the end of the URL. With or without, the actual keywords in the URL are more important.
The use of rel="canonical" has a very different purpose. Say for whatever reason you want to keep both of your URL's even though they have the exact same content (testing conversation rates, for example). In this case you would use a rel="canonical" on the page you don't want to rank in the search engines, pointing to the page you do want to rank for.
On http://www.mysite/blue/index.html for example, you would create this tag: <rel="canonical" href="http://www.mysite/blue/">eCommerce sites have to do this a lot.</rel="canonical">
Rel canonical should not be used when you're trying to move content from one URL to another. That's what 301s are for.
-
If you are only talking about your home page, then yes setup a 301 redirect as others have shown for the entries already in googles index BUT a redirect itself can lose up to 10% of any link juice flowing to your index page. And if you're building off site links, do you link to your root domain or the specific URL of your homepage? My guess is the root, i.e. www.mysite.com so unless www.mysite.com is actually a different website to whats found at www.mysite.com/blue/ then I always strive to get my sites working without an initial redirect taking place when someone goes to www.mysite.com
Depending on you choice of webserver, you can specify what the default index page should be, in apache this is known as the 'DirectoryIndex'.
If you add the line
DirectoryIndex /blue/index.html
to your .htaccess (or even better apache site config if you can) then apache will serve that page WITHOUT the redirect ensuring any link juice to your route domain is not diluted.
Then just make sure any links on your own site that point to you home page DO NOT point to /blue/ or /blue/index.html but simply to "/" or "http://www.mysite.com/"
-
Hi Perri,
This is an older thread, and people may not see the new response if they're not subscribed to it.
You can certainly redirect the index.html to /. The above thread gives some help, as well as http://www.smartlabsoftware.com/howto/redirect-index-page.htm (though I don't know the age of that post and if it's for a current version of Apache).
I suggest opening a new question here with a title something like "redirecting /index.html to / in apache" and give your details in that question, with a link to the above URL and ask if this is still valid. A link to your site in the question, if you can give it, would also be great.
Thanks!
-
A while back I had asked our hosting company to create 301 redirects in the htacces file for the same issue (www.mysite.com/index.html to www.mysite.com, www.mysite.com/products/index.html to www.mysite.com/products/ .....) THe response I received was "redirecting .../index.html to ..../ won’t work. They’re the same page. Apache will get in an infinite loop and the page won’t load. "
Any help would be greatly appreciated since I have 36 instances of this happening on our site.
-
Hi Norberto,
Are you still having duplicate content errors, or did you clear this up? We're happy to help if you're still having any problems, just add a response to this thread.
-
Hi guys thank you so much for your help. I have another think, looking deep in the report i saw a duplicate page title, it is a section like products reviews, for example like this:
Url 1: Title 1 : I like the product - product reviw name of product
Url 2: Title 2 : How can i get the product - product reviw name of product
Why this is condider like a same title page, can somebody help me to clear this doubt.
And what can i do to resolve this problem.
Thanks
-
The only place you would see something in WMT would be DIAGNOSTICS > HTML SUGGESTIONS, and it should show as duplicate title tag and meta description.
WMT wouldn't flag that up specifically because it's not something it really checks for.
-
Thanks for your help guys.
I have another question, why in google websmaster tools i don't recieve any massage error about it?
-
Sorry Saibose, disagree entirely, canonical is a band aid whereas 301 is a fix.
-
Something very similar came up earlier, best to 301 them as E-Dreamz says.
If you put this in your .htaccess file you should have all your pages as www. and the index.html will disappear.
RewriteEngine On Options +FollowSymLinks
RewriteCond %{HTTP_HOST} ^example.com
RewriteRule (.*) http://www.example.com/$1 [R=301,L]RewriteCond %{THE_REQUEST} ^./index.html
RewriteRule ^(.)index.html$ http://www.example.com/$1 [R=301,L] -
Add a rel=canonical tag for the pages that have multiple URLs.
You can see some resources here.
http://www.google.com/support/webmasters/bin/answer.py?hl=en&answer=139394
If you use the rel= canonical tag you will have to do it for all pages with this issue. If the issue is widespread, you can consider a 301, but that wont be very effective for SEO purposes when compared to rel=canonical tag.
-
I would recommend a 301 redirect.
For the best SEO value you want to leave off the /index.html - espcially if that is your homepage.
Don't forget to redirect non-www to www.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate Content
I am trying to get a handle on how to fix and control a large amount of duplicate content I keep getting on my Moz Reports. The main area where this comes up is for duplicate page content and duplicate title tags ... thousands of them. I partially understand the source of the problem. My site mixes free content with content that requires a login. I think if I were to change my crawl settings to eliminate the login and index the paid content it would lower the quantity of duplicate pages and help me identify the true duplicate pages because a large number of duplicates occur at the site login. Unfortunately, it's not simple in my case because last year I encountered a problem when migrating my archives into a new CMS. The app in the CMS that migrated the data caused a large amount of data truncation Which means that I am piecing together my archives of approximately 5,000 articles. It also means that much of the piecing together process requires me to keep the former app that manages the articles to find where certain articles were truncated and to copy the text that followed the truncation and complete the articles. So far, I have restored about half of the archives which is time-consuming tedious work. My question is if anyone knows a more efficient way of identifying and editing duplicate pages and title tags?
Technical SEO | | Prop650 -
Does adding a noindex tag reduce duplicate content?
I've been working under the assumption for some time that if I have two (or more) pages which are very similar that I can add a noindex tag to the pages I don't need and that will reduce duplicate content. As far as I know this removes the pages with the tag from Google's index and stops any potential issues with duplicate content. It's the second part of that assumption that i'm now questioning. Despite pages having the noindex tag they continue to appear in Google Search console as duplicate content, soft 404 etc. That is, new pages are appearing regularly that I know to have the noindex tag. My thoughts on this so far are that Google can still crawl these pages (although won't index them) so shows them in GSC due to a crude issue flagging process. I mainly want to know: a) Is the actual Google algorithm sophisticated enough to ignore these pages even through GSC doesn't. b) How do I explain this to a client.
Technical SEO | | ChrisJFoster0 -
Over 700+ duplicate content pages -- help!
I just signed up for SEO Moz pro for my site. The initial report came back with over 700+ duplicate content pages. My problem is that while I can see why some of the content is duplicated on some of the pages I have no idea why it's coming back as duplicated. Is there a tutorial for a novie on how to read the duplicate content report and what steps to take? It's an e-commerce website and there is some repetitive content on all the product pages like our "satisfaction guaranteed" text and the fabric material... and not much other text. There's not a unique product description because an image speaks for itself. Could this be causing the problem? I have lots of URLs with over 50+ duplicates. Thx for any help.
Technical SEO | | Santaur0 -
Avoiding duplicate content on product pages?
Hi, I'm creating a bunch of product pages for courses for a university and I'm concerned about duplicate content penalties. While the page names are different and some of the test is different, much of the text is the same between pairs of pages. I.e. a BA and an MA in a particular subject (say 'hairdressing' will have the same subject descriptions, school introduction paragraph, industry overview paragraph etc. 1. Is this a problem? In a site with 100 pages, if sets of 2 pages have about 50% identical content... 2. If it is a problem, is there anything I can do, other than rewrite the text? 3. From a search perspective, would both pages show up in search results in searches related to 'hairdressing courses' 'study hairdressing' etc? Thanks!
Technical SEO | | AISFM0 -
Duplicate Content on Product Pages
Hello I'm currently working on two sites and I had some general question's about duplicate content. For the first one each page is a different location, but the wording is identical on each; ie it says Instant Remote Support for Critical Issues, Same Day Onsite Support with a 3-4 hour response time, etc. Would I get penalized for this? Another question i have is, we offer Antivirus support for providers ie Norton, AVG,Bit Defender etc. I was wondering if we will get penalized for having the same first paragraph with only changing the name of the virus provider on each page? My last question is we provide services for multiple city's and towns in various states. Will I get penalized for having the same content on each page, such as towns and producuts and services we provide? Thanks.
Technical SEO | | ilyaelbert0 -
Duplicate content by category name change
Hello friends, I have several problems with my website related with duplicate content. When we changed any family name, for example "biodiversidad" to "cajas nido y biodiversidad", it creates a duplicate content because: mydomain.com/biodiversidad and mydomain.com/cajas-nido-y-biodiversidad have the same content. This happens every tame I change the names of the categories or families. To avoid this, the first thing that comes to my mid is a 301 redirect from the old to the new url, but I wonder if this can be done more automatically otherwise, maybe a script? Any suggestion? Thank you
Technical SEO | | pasape0 -
Duplicate Content
Hi - We are due to launch a .com version of our site, with the ability to put prices into local currency, whereas our .co.uk site will be solely £. If the content on both the .com and .co.uk sites is the same (at product level mainly), will we be penalised? What is the best way to get around this?
Technical SEO | | swgolf1230 -
The Bible and Duplicate Content
We have our complete set of scriptures online, including the Bible at http://lds.org/scriptures. Users can browse to any of the volumes of scriptures. We've improved the user experience by allowing users to link to specific verses in context which will scroll to and highlight the linked verse. However, this creates a significant amount of duplicate content. For example, these links: http://lds.org/scriptures/nt/james/1.5 http://lds.org/scriptures/nt/james/1.5-10 http://lds.org/scriptures/nt/james/1 All of those will link to the same chapter in the book of James, yet the first two will highlight the verse 5 and verses 5-10 respectively. This is a good user experience because in other sections of our site and on blogs throughout the world webmasters link to specific verses so the reader can see the verse in context of the rest of the chapter. Another bible site has separate html pages for each verse individually and tends to outrank us because of this (and possibly some other reasons) for long tail chapter/verse queries. However, our tests indicated that the current version is preferred by users. We have a sitemap ready to publish which includes a URL for every chapter/verse. We hope this will improve indexing of some of the more popular verses. However, Googlebot is going to see some duplicate content as it crawls that sitemap! So the question is: is the sitemap a good idea realizing that we can't revert back to including each chapter/verse on its own unique page? We are also going to recommend that we create unique titles for each of the verses and pass a portion of the text from the verse into the meta description. Will this perhaps be enough to satisfy Googlebot that the pages are in fact unique? They certainly are from a user perspective. Thanks all for taking the time!
Technical SEO | | LDS-SEO0