Duplicate content pages on different domains, best practice?
-
Hi,
We are running directory sites on different domains of different countries (we have the country name in the domain name of each site) and we have the same static page on each one, well, we have more of them but I would like to exemplify one static page for the sake of simplicity.
So we have http://firstcountry.com/faq.html, http://secondcountry.com/faq.html and so on for 6-7 sites, faq.html from one country and the other have 94% similarity when checked against duplicate content. We would like an alternative approach to canonical cause the content couldn´t belong to only one of this sites, it belongs to all. Second option would be unindex all but one country. It´s syndicated content but we cannot link back to the source cause there is none.
Thanks for taking the time in reading this.
-
Using canonical IS NOT the solution, because if you use canonical the FAQ pages of the canonicalized websites are going to be deindexed.
So, just do it if you really don't care about the traffic those answers can generate for your sites (as you can imagine, this is an ironic suggestion...).
Just use the hreflang, because Google in the last months has become quite smart in understanding that it means that you consider those pages relevant enough for the geo-targeted audiences to not filter them out even if they are substantially identical between country versions.
Said that, try to differentiate the FAQ pages (better localization of the language. i.e.: English UK is slightly different than American English), or even offering a local number for phone inquiries and localized email address for question via email.
In general, it is not a good idea using the crossdomain canonical in International SEO, and it should be used just in exceptional cases.
-
In order to make things easier you can implement hreflang via sitemaps.xml using this tool by Mediaflow: http://www.themediaflow.com/tool_hreflang.php.
-
If your site is based on templates so you can easily put in a header code (Wordpress, Joomla, most CMS, anything but a page-by-page HTML usually) you can insert it directly in by calling in the page itself like this:
" hreflang="x-default" />
" hreflang="en-au" />
" hreflang="en-us" />
" hreflang="en-nz" />This works on Apache servers - this starts with the domain and then request_URI pulls in the page you're on so /about, or /faq and adds the appropriate hreflang tag to that.
Also, when you're done implementing hreflang test it using Flang.
-
As the other users have pointed out, the alternate and hreflang tag would be most ideal. I am in a pickle myself with a very similar issue.
You must note that the alternate tag is to be applied on a page level so every page should resolve to the appropriate URL of it's copy on all other country domains.
So your homepage (.com) could have the following alternate tags:
But on your FAQ page, the alternates would be:
You'll have to rinse and repeat on all 3 sites and for every single page.
Tedious if you ask me! Does anyone know an easier way to go around adding alternate tags to 3 or 4 sites without doing it manually?
The advantage of implementing those however is that you are not canonicalising to one domain which means all your domains stand a chance of performing well in their regions (e.g a search on Google Australia will show the .com.au website).
Again, does anyone have a better approach to this or seen / heard of one? Apart from canonical of course.
-
Hreflang tags are great. I would highly suggest implementing these. Something that I was confused about when I first started using them was that all tags should be on all domains including its own.
For example: firstcountry.com/faq.html should have tags for:
and so on.
You can check that these have been implemented correctly in Google Webmaster Tools under "Search Traffic" -> "International Targeting"
-
I would start by implementing hreflang tags:
https://support.google.com/webmasters/answer/189077?hl=en
Hreflang should take care of these type of issues as Google will associate the right country domain with the content. You may see some overlap for awhile - we've seen hreflang take a bit longer than we'd like to get fully set but once it is, it usually works well.
Short of that, you have 3 options. 1) change the content on all sites to be (somewhat) unique. 2) deindex all but one as you said, 3) canonical, as you said.
1, 2 & 3 all have problems so that's why I would start with hreflang.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate content on report
Hi, I just had my Moz Campaign scan 10K pages out of which 2K were duplicate content and URL's are http://www.Somesite.com/modal/register?destination=question%2F37201 http://www.Somesite.com/modal/register?destination=question%2F37490 And the title for all 2K is "Register" How can i deal with this as all my pages have the register link and login and when done it comes back to the same page where we left and that it actually not duplicate but we need to deal with it propely thanks
Technical SEO | | mtthompsons0 -
Duplicate Page Content
Hi, I just had my site crawled by the seomoz robot and it came back with some errors. Basically it seems the categories and dates are not crawling directly. I'm a SEO newbie here Below is a capture of the video of what I am talking about. Any ideas on how to fix this? Hkpekchp
Technical SEO | | mcardenal0 -
How to prevent duplicat content issue and indexing sub domain [ CDN sub domain]?
Hello! I wish to use CDN server to optimize my page loading time ( MaxCDN). I have to use a custom CDN sub domain to use these services. If I added a sub domain, then my blog has two URL (http://www.example.com and http://cdn.example.com) for the same content. I have more than 450 blog posts. I think it will cause duplicate content issues. In this situation, what is the best method (rel=canonical or no-indexing) to prevent duplicate content issue and prevent indexing sub domain? And take the optimum service of the CDN. Thanks!
Technical SEO | | Godad0 -
Duplicate Page Content for sorted archives?
Experienced backend dev, but SEO newbie here 🙂 When SEOmoz crawls my site, I get notified of DPC errors on some list/archive sorted pages (appending ?sort=X to the url). The pages all have rel=canonical to the archive home. Some of the pages are shorter (have only one or two entries). Is there a way to resolve this error? Perhaps add rel=nofollow to the sorting menu? Or perhaps find a method that utilizes a non-link navigation method to sort / switch sorted pages? No issues with duplicate content are showing up on google webmaster tools. Thanks for your help!
Technical SEO | | jwondrusch0 -
Duplicates on the page
Hello SEOMOZ, I've one big question about one project. We have a page http://eb5info.com/eb5-attorneys and a lot of other similar pages. And we got a big list of errors, warnings saying that we have duplicate pages. But in real not all of them are same, they have small differences. For example - you select "State" in the left sidebar and you see a list on the right. List on the right panel is changing depending on the what you selecting on the left. But on report pages marked as duplicates. Maybe you can give some advices how to improve quality of the pages and make SEO better? Thanks Igor
Technical SEO | | usadvisors0 -
How do I fix this type of duplicate page content problem?
Sample URLs with this Duplicate Page Content URLs Internal Links External Links Page Authority Linking Root Domains http://rogerelkindlaw.com/index.html 30 0 26 1 http://www.rogerelkindlaw.com/index.html 30 0 20 1 http://www.rogerelkindlaw.com/ | 1,630 | 613 | 43 | 110 | As you can see there are three duplicate pages; http://rogerelkindlaw.com/index.html http://www.rogerelkindlaw.com/index.html http://www.rogerelkindlaw.com/ What would be the best and most efficient way to fix this problem and also how to prevent this from happening? Thank you.
Technical SEO | | brianhughes0 -
Is 100% duplicate content always duplicate?
Bit of a strange question here that would be keen on getting the opinions of others on. Let's say we have a web page which is 1000 lines line, pulling content from 5 websites (the content itself is duplicate, say rss headlines, for example). Obviously any content on it's own will be viewed by Google as being duplicate and so will suffer for it. However, given one of the ways duplicate content is considered is a page being x% the same as another page, be it your own site or someone elses. In the case of our duplicate page, while 100% of the content is duplicate, the page is no more than 20% identical to another page so would it technically be picked up as duplicate. Hope that makes sense? My reason for asking is I want to pull latest tweets, news and rss from leading sites onto a site I am developing. Obviously the site will have it's own content too but also want to pull in external.
Technical SEO | | Grumpy_Carl0 -
Up to my you-know-what in duplicate content
Working on a forum site that has multiple versions of the URL indexed. The WWW version is a top 3 and 5 contender in the google results for the domain keyword. All versions of the forum have the same PR, but but the non-WWW version has 3,400 pages indexed in google, and the WWW has 2,100. Even worse yet, there's a completely seperate domain (PR4) that has the forum as a subdomain with 2,700 pages indexed in google. The dupe content gets completely overwhelming to think about when it comes to the PR4 domain, so I'll just ask what you think I should do with the forum. Get rid of the subdomain version, and sometimes link between two obviously related sites or get rid of the highly targeted keyword domain? Also what's better, having the targeted keyword on the front of Google with only 2,100 indexed pages or having lower rankings with 3,400 indexed pages? Thanks.
Technical SEO | | Hondaspeder0