Why am I getting all these duplicate pages?
-
This is going for basically all my pages, but my website has 3 'duplicates' as the rest just have 2 (no index)
Why are these 3 variations counting as duplicate pages?
-
Actually, canonical tags are the absolute last-ditch way of dealing with this issue.
The correct solution is to use 301-redirects to force all version of the URL except the primary to redirect to the primary (also called canonical) URL. Canonical in this instance just means the primary or most authoritative version of something. Nothing to do with the tags of the same name.
The only reason to use the rel=canonical tag for this is if you have absolutely no way to do it through 301-redirects. (For instance your host doesn't allow access to the .htaccess file and your DNS system doesn't allow it either.)
Use Travis's info below for exactly how to do this in .htaccess. There are also many other posts here in Q&A that address this if you want more reference points.
Paul
-
Your next question is; "Great, but how do I fix it?"
It looks like this particular detail was missed during server configuration. You would handle this with rewrites via .htaccess if you're using an Apache server. However, if you're unfamiliar with the file - proceed with caution - if you can't push and pull from a test environment for some reason. A little bit of white space or a syntax error can knock the site down until you find the error.
Otherwise, Ultimate Htaccess has just about everything you need to know. Here are the commands you will need. If you're using WordPress, make sure the redirects go before the section of your file that pertains to WordPress.
-
Technically all of these urls are different. A web server could return completely different content for all the urls above. When Google “canonicalizes” a url, pick the url that seems like the best representative from that set.
Check this link: http://moz.com/learn/seo/duplicate-content
-
Hello W2GITeam and welcome to the world of SEO!
The problem you've described is covered in basic/fundamental SEO concepts. The specific topic to help you turn those 3 pages into a single, non-duplicate, indexed page is through Canonical tags.
Learn more about those tags and how they help suppress duplicate content from being indexed here: http://moz.com/learn/seo/canonicalization
That's plenty to get you started
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Joomla Opencart Duplicate Link Issue
With a few recent catalog changes my client is having tons of duplicate content issues. The most peculiar is one that I cannot see visibly on the site pages. This page: http://www.CLIENTDOMAIN.com/catalog/lights.html shows to have several duplicate content issues with URLs such as: http://www.CLIENTDOMAIN.com/catalog/lights/Lamp-DRW.htmlhttp:/www.CLIENTDOMAIN.com - where the domain is basically added to the end of the first URL. Has anyone seen this? Any fixes, plugins or thoughts? Thanks,
Moz Pro | | KristinaWitmer
Kristina0 -
How to increase page authority
I wonder how to increase the page authority or the domain authority to begin with. It seems you are putting a lot of weight on this in your analysis.
Moz Pro | | wcsinc0 -
Rogerbot getting cheeky?
Hi SeoMoz, From time to time my server crashes during Rogerbot's crawling escapades, even though I have a robots.txt file with a crawl-delay 10, now just increased to 20. I looked at the Apache log and noticed Roger hitting me from from 4 different addresses 216.244.72.3, 72.11, 72.12 and 216.176.191.201, and most times whilst on each separate address, it was 10 seconds apart, ALL 4 addresses would hit 4 different pages simultaneously (example 2). At other times, it wasn't respecting robots.txt at all (see example 1 below). I wouldn't call this situation 'respecting the crawl-delay' entry in robots.txt as other question answered here by you have stated. 4 simultaneous page requests within 1 sec from Rogerbot is not what should be happening IMHO. example 1
Moz Pro | | BM7
216.244.72.12 - - [05/Sep/2012:15:54:27 +1000] "GET /store/product-info.php?mypage1.html" 200 77813
216.244.72.12 - - [05/Sep/2012:15:54:27 +1000] "GET /store/product-info.php?mypage2.html HTTP/1.1" 200 74058
216.244.72.12 - - [05/Sep/2012:15:54:28 +1000] "GET /store/product-info.php?mypage3.html HTTP/1.1" 200 69772
216.244.72.12 - - [05/Sep/2012:15:54:37 +1000] "GET /store/product-info.php?mypage4.html HTTP/1.1" 200 82441 example 2
216.244.72.12 - - [05/Sep/2012:15:46:15 +1000] "GET /store/mypage1.html HTTP/1.1" 200 70209
216.244.72.11 - - [05/Sep/2012:15:46:15 +1000] "GET /store/mypage2.html HTTP/1.1" 200 82384
216.244.72.12 - - [05/Sep/2012:15:46:15 +1000] "GET /store/mypage3.html HTTP/1.1" 200 83683
216.244.72.3 - - [05/Sep/2012:15:46:15 +1000] "GET /store/mypage4.html HTTP/1.1" 200 82431
216.244.72.3 - - [05/Sep/2012:15:46:16 +1000] "GET /store/mypage5.html HTTP/1.1" 200 82855
216.176.191.201 - - [05/Sep/2012:15:46:26 +1000] "GET /store/mypage6.html HTTP/1.1" 200 75659 Please advise.1 -
Need to find all pages that link to list of pages/pdf's
I know I can do this in OSE page by page, but is there a way I can do this in a large batch? There are 200+ PDF's that I need to figure out what pages (if any) link to the PDF. I'd rather not do this page by page, but rather copy-paste the entire list of pages I'm looking for. Any tools you know of that can do this?
Moz Pro | | ryanwats0 -
Duplicate Page Titles and Content
The SeoMoz crawler has found many pages like this on my site with /?Letter=Letter, e.g. http://www.johnsearles.com/metal-art-tiles/?D=A. I believe it is finding multiple caches of a page and identifying them as duplicates. Is there any way to screen out these multiple cache results?
Moz Pro | | johnsearles0 -
How To Solve Too Many On-Page Links In Blogger?
Hi, I Have An Issue Too Many On-Page Links In My Site And I Saw That There Are More Than 300 On Page Links On My Home Page URL. My Site Is Hosted On Blogger. So Please Tell Me How To Fix This Problem In Blogger.
Moz Pro | | MaherHackers0 -
Broken Links and Duplicate Content Errors?
Hello everybody, I’m new to SEOmoz and I have a few quick questions regarding my error reports: In the past, I have used IIS as a tool to uncover broken links and it has revealed a large amount of varying types of "broken links" on our sites. For example, some of them were links on my site that went to external sites that were no longer available, others were missing images in my CSS and JS files. According to my campaign in SEOmoz, however, my site has zero broken links (4XX). Can anyone tell me why the IIS errors don’t show up in my SEOmoz report, and which of these two reports I should really be concerned about (for SEO purposes)? 2. Also in the "errors" section, I have many duplicate page titles and duplicate page content errors. Many of these "duplicate" content reports are actually showing the same page more than once. For example, the report says that "http://www.cylc.org/" has the same content as "http://www.cylc.org/index.cfm" and that, of course, is because they are the same page. What is the best practice for handling these duplicate errors--can anyone recommend an easy fix for this?
Moz Pro | | EnvisionEMI0 -
How do I get the SERP overlay tool to work?
I have the SEOMoz toolbar installed. In the settings I have a tick next to Display SERP Overlay. When I first activated this it showed up but with no data just continually searching for a long time. Now it is not showing up at all. This is such a great tool, how can I get it working? Thanks, Daniel.
Moz Pro | | iSenseWebSolutions0