Duplicate page content and Duplicate page title errors
-
Hi,
I'm new to SeoMoz and to this forum. I've started a new campaign on my site and got back loads of error.
Most of them are Duplicate page content and Duplicate page title errors. I know I have some duplicate titles but I don't have any duplicate content.
I'm not a web developer and not so expert but I have the impression that the crawler is following all my internal links (Infact I have also plenty of warnings saying "Too many on-page links".
Do you think this is the cause of my errors? Should I implement the nofollow on all internal links? I'm working with Joomla.
Thanks a lot for your help
Marco
-
Hi Marco,
I took a look at your page at http://www.beautifulpuglia.com/it/linea-costiera/isole-tremiti.html
Looks like you've got the canonical in place okay here. The next step is to add the canonical on every page that is a duplicate of this page. And you want to make sure to point to the right page. Let me be clear: Every page that is a duplicate of this page should have the same canonical. In this case:
<link rel=”canonical” href=[”http://www.beautifulpuglia.com/it/gargano/isole-tremiti.html”/](view-source:http://www.beautifulpuglia.com/linea-costiera/%E2%80%9Dhttp://www.beautifulpuglia.com/it/gargano/isole-tremiti.html%E2%80%9D/)>
You can find the other pages you need to add this tag to in your SEOmoz report. In each duplicated content report, it will list the number of other pages that are duplicates. Simply click on the number to see the URLs.
I'm not a Joomla expert, but webmasters I've talked to have expressed that other platforms such as Wordpress and Drupal are much more accommodating of these types of fixes. There are some various plugin modules you can use, but you'll have to select one appropriate to your configuration.
Here's a good resource from Dr. Pete: http://www.seomoz.org/blog/duplicate-content-in-a-post-panda-world
Hope this helps. Best of luck.
-
Elias,
I too have 'thousands' of duplicated errors in SEOmoz. Most of which are because it is returning
/abc.com as a different page to /ABC.com
Surely Google doesn't do that? Just because one URL is in capital and the other small case? I also have no idea where SEOmoz is picking that up from......possibly links internal to the page with the hyperlink using different case?
It seems to me this is too sensitive and for me to fix that would take WEEKS!!!! I fail to see if there would be any uplift if Google sees beyond that issue as its cosmetic and not functional.
Regards
Andy
-
It looks fine to me. You will need to do the same on all of your pages.
If you've just added the code you will need to wait up to a week for SEOmoz to re-crawl your website depending on when you're site crawl is scheduled.
Let me know how you get on.
Elias
-
Hi Elias, Hi Marisa,
thanks you both
you are right, in the meantime I had done this but I have the impression it is not working and I don't know what I'm doing wrong.
I'm attaching a link to a page of my site (I hope I can do this). Please have a look at the code, you will see the tag rel=”canonical” href=”http://www.beautifulpuglia.com/it/gargano/isole-tremiti.html”/> which is indicating the URL I want to use. However SeoMoz is still giving me the error. And this is happening for both the Italian and English version.
So far I've only added the tag to this page, I want to find the solution before modifying all pages currently affected.
http://www.beautifulpuglia.com/it/linea-costiera/isole-tremiti.html
Thanks a lot again
-
Hi Marco, as Marissa says - by putting the canonical tag on one page you are putting it on all of them as they are in fact the same page - they are just reached by different URLs.
-
www.site.com/ and www.site.com/index.html, site.com/index.html/, ect, are already the same page. So, there's only one page TO put the tag on. You're just telling the crawlers that you only want one of them to get the credit, and which version of the page you prefer to be displayed.
-
Hi Elias,
thanks a lot for your reply. I've read few posts about the canonical tag and Yes I'm going to try it.
Just couple of things:
-
Let's say I have 4 duplicate for one page, I presume I have to add the tag in the head of only one page right? Does it make any difference which one I pick?
-
Any idea on how this can be implemented in Joomla?It doesn't seem to be very straightforward.
Thanks a lot
Marco
-
-
Hi Marco,
It seems to me like you need to implement the canonical tag.
Site crawlers/bots will consider the following pages as different pages because of their URL and thus tell indicate to them that the content is duplicated on each page...
By implementing the following tag on each of your sites pages (changing the URL for each page) you will tell the crawler which page they should be indexing and to ignore the other.
Here's an example of a canonical tag (to be placed within the head tag of the page)
I think this will sort out your duplication issues.
You can find more information about canonical URLs here http://www.seomoz.org/blog/canonical-url-tag-the-most-important-advancement-in-seo-practices-since-sitemaps
I hope this helps!
Elias
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Content Strategy/Duplicate Content Issue, rel=canonical question
Hi Mozzers: We have a client who regularly pays to have high-quality content produced for their company blog. When I say 'high quality' I mean 1000 - 2000 word posts written to a technical audience by a lawyer. We recently found out that, prior to the content going on their blog, they're shipping it off to two syndication sites, both of which slap rel=canonical on them. By the time the content makes it to the blog, it has probably appeared in two other places. What are some thoughts about how 'awful' a practice this is? Of course, I'm arguing to them that the ranking of the content on their blog is bound to be suffering and that, at least, they should post to their own site first and, if at all, only post to other sites several weeks out. Does anyone have deeper thinking about this?
Intermediate & Advanced SEO | | Daaveey0 -
Duplicate Content through 'Gclid'
Hello, We've had the known problem of duplicate content through the gclid parameter caused by Google Adwords. As per Google's recommendation - we added the canonical tag to every page on our site so when the bot came to each page they would go 'Ah-ha, this is the original page'. We also added the paramter to the URL parameters in Google Wemaster Tools. However, now it seems as though a canonical is automatically been given to these newly created gclid pages; below https://www.google.com.au/search?espv=2&q=site%3Awww.mypetwarehouse.com.au+inurl%3Agclid&oq=site%3A&gs_l=serp.3.0.35i39l2j0i67l4j0i10j0i67j0j0i131.58677.61871.0.63823.11.8.3.0.0.0.208.930.0j3j2.5.0....0...1c.1.64.serp..8.3.419.nUJod6dYZmI Therefore these new pages are now being indexed, causing duplicate content. Does anyone have any idea about what to do in this situation? Thanks, Stephen.
Intermediate & Advanced SEO | | MyPetWarehouse0 -
Duplicate pages with http and https
Hi all, We changed the payment part of our site to https from http a while ago. However once on the https pages, all the footer and header links are relative URLs, so once users have reached the payment pages and then re-navigate back to other pages in our website they stay on https. The build up of this happening has led to Google indexing all our pages in https (something we did not want to happen), and now we are in the situation where our homepage listing on Google is https rather than http. We would prefer the organic listings to be http (rather than https) and having read lots on this (included the great posts on the moz (still feels odd not refering to it as seomoz!) blog around this subject), possible solutions include redirects or a canoncial tags. My additional questions around these options are: 1. We already have 2 redirects on some pages (long story), will another one negatively impact our rankings? 2. Is a canonical a strong enough hint to Google to stop Google indexing the https versions of these page to the extent that out http pages will appear in natural listings again? If anyone has any other suggestions or other ideas of how to address this issue, that would be great! Thanks 🙂 Diana
Intermediate & Advanced SEO | | Diana.varbanescu0 -
International SEO - cannibalisation and duplicate content
Hello all, I look after (in house) 3 domains for one niche travel business across three TLDs: .com .com.au and co.uk and a fourth domain on a co.nz TLD which was recently removed from Googles index. Symptoms: For the past 12 months we have been experiencing canibalisation in the SERPs (namely .com.au being rendered in .com) and Panda related ranking devaluations between our .com site and com.au site. Around 12 months ago the .com TLD was hit hard (80% drop in target KWs) by Panda (probably) and we began to action the below changes. Around 6 weeks ago our .com TLD saw big overnight increases in rankings (to date a 70% averaged increase). However, almost to the same percentage we saw in the .com TLD we suffered significant drops in our .com.au rankings. Basically Google seemed to switch its attention from .com TLD to the .com.au TLD. Note: Each TLD is over 6 years old, we've never proactively gone after links (Penguin) and have always aimed for quality in an often spammy industry. **Have done: ** Adding HREF LANG markup to all pages on all domain Each TLD uses local vernacular e.g for the .com site is American Each TLD has pricing in the regional currency Each TLD has details of the respective local offices, the copy references the lacation, we have significant press coverage in each country like The Guardian for our .co.uk site and Sydney Morning Herlad for our Australia site Targeting each site to its respective market in WMT Each TLDs core-pages (within 3 clicks of the primary nav) are 100% unique We're continuing to re-write and publish unique content to each TLD on a weekly basis As the .co.nz site drove such little traffic re-wrting we added no-idex and the TLD has almost compelte dissapread (16% of pages remain) from the SERPs. XML sitemaps Google + profile for each TLD **Have not done: ** Hosted each TLD on a local server Around 600 pages per TLD are duplicated across all TLDs (roughly 50% of all content). These are way down the IA but still duplicated. Images/video sources from local servers Added address and contact details using SCHEMA markup Any help, advice or just validation on this subject would be appreciated! Kian
Intermediate & Advanced SEO | | team_tic1 -
PDF for link building - avoiding duplicate content
Hello, We've got an article that we're turning into a PDF. Both the article and the PDF will be on our site. This PDF is a good, thorough piece of content on how to choose a product. We're going to strip out all of the links to our in the article and create this PDF so that it will be good for people to reference and even print. Then we're going to do link building through outreach since people will find the article and PDF useful. My question is, how do I use rel="canonical" to make sure that the article and PDF aren't duplicate content? Thanks.
Intermediate & Advanced SEO | | BobGW0 -
Duplicate content question? thanks
Hi, Im my time as an SEO I have never come across the following two scenarios, I am an advocate of using unique content, therefore always suggest and in cases demand that all content is written or re-written. This is the scenarios I am facing right now. For Example we have www.abc.com (has over 200 original recipes) and then we have www.xyz.com with the recipes but they are translated into another language as they are targeting different audiences, will Google penalize for duplicate content? The other issue is that the client got the recipes from www.abc.com (that have been translated) and use them in www.xyz.com aswell, both sites owned by the same company so its not pleagurism they have legal rights but I am not sure how Google will see it and if it will penalize the sites. Thanks!
Intermediate & Advanced SEO | | M_81 -
SEOMOZ duplicate page result: True or false?
SEOMOZ say's: I have six (6) duplicate pages. Duplicate content tool checker say's (0) On the physical computer that hosts the website the page exists as one file. The casing of the file is irrelevant to the host machine, it wouldn't allow 2 files of the same name in the same directory. To reenforce this point, you can access said file by camel-casing the URI in any fashion (eg; http://www.agi-automation.com/Pneumatic-grippers.htm). This does not bring up a different file each time, the server merely processes the URI as case-less and pulls the file by it's name. What is happening in the example given is that some sort of indexer is being used to create a "dummy" reference of all the site files. Since the indexer doesn't have file access to the server, it does this by link crawling instead of reading files. It is the crawler that is making an assumption that the different casings of the pages are in fact different files. Perhaps there is a setting in the indexer to ignore casing. So the indexer is thinking that these are 2 different pages when they really aren't. This makes all of the other points moot, though they would certainly be relevant in the case of an actual duplicated page." ****Page Authority Linking Root Domains http://www.agi-automation.com/ 43 82 http://www.agi-automation.com/index.html 25 2 http://www.agi-automation.com/Linear-escapements.htm 21 1 www.agi-automation.com/linear-escapements.htm 16 1 http://www.agi-automation.com/Pneumatic-grippers.htm 30 3 http://www.agi-automation.com/pneumatic-grippers.htm 16 1**** Duplicate content tool estimates the following: www and non-www header response; Google cache check; Similarity check; Default page check; 404 header response; PageRank dispersion check (i.e. if www and non-www versions have different PR).
Intermediate & Advanced SEO | | AGIAutomation0 -
Odd duplicate page notification (I think)...
Hi! In looking at my site's crawl diagnostics, I came across 2 pages that were flagged as duplicate content. I can't quite figure out why. The only difference in the URLS is an uppercase '"B" vs a lowercase "b" following the "~". Here are the URLS: lowercase b example:
Intermediate & Advanced SEO | | peterdbaron
http://www.admissionsquest.com/~boardingSchoolNotes/ShowArticle.cfm/ArticleID/142/ArticleTypeID/12/Topic/What-Makes-a-Progressive-Boarding-School uppercase B example:
http://www.admissionsquest.com/~BoardingSchoolNotes/ShowArticle.cfm/ArticleID/142/ArticleTypeID/12/Topic/What-Makes-a-Progressive-Boarding-School Is that the problem? Any advice is very much appreciated. Thanks! Peter0