Geotargeting duplicate content to different regions - href and canonical tag confusion
-
If you duplicate content onto a sub-folder for say a new US geotargeted site (to target kw spelling differences) and, in addition to GWT geotargeting settings, implement the 'Canonical' and 'Hreflang' tags on these new pages to show G different region and language version (en-us). Then does the original/main site similar pages also need to have canonical and href tags ?
The main/original sites page I don't really want to target a specific country (although existing signals (hosting etc) will be UK (primary target of main site) but pages show up in other country searches too (which we want).
Im presuming fine to leave the original/main site as it currently is although wording in google blog/webmaster central articles etc are a bit confusing hence why im asking for anyone elses opinion/input on this.
Also is there are any benefit (or just best practice) to use 'www.example.com/en-us/...' in the subdirectory URL as opposed to just 'www.example.com/us/'
many thanks in advance to any commentators
-
Many thanks Gianluca !!
-
Hi,
I suggest you both to give a read to this post by DejanSEO, which is quite clear and - IMHO - points to the right interpretation of a somehow confused best practice.
-
Thats what i thought originally but getting confised when i read this page: http://googlewebmastercentral.blogspot.com.au/2011/12/new-markup-for-multilingual-content.html
Specifically this bit:
Annotating pages as substantially similar content
Optionally, for pages that have substantially the same content in the same language and are targeted at multiple countries, you may use the rel="canonical" link element to specify your preferred version. We’ll use that signal to focus on that version in search, while showing the local URLs to users where appropriate. For example, you could use this if you have the same product page in German, but want to target it separately to users searching on the Google properties for Germany, Austria, and Switzerland.
And read in conjunction with this article:
Specifically this bit:
The Effect Of Combining Canonical Tags & Hreflang Tags
Not forgetting that the canonical tags should only be used with content in the same language, when would we use both?
Well firstly, the use of both would involve what I usually call world languages such as English, Spanish, French or Portuguese. These languages are used in many countries and, whilst there are variations between the use of these languages in those countries, the variations are sometimes small.
Additionally, multinational publishers often save costs by using one version of the language for all countries speaking that general language, thus ignoring the regional variations. In other words, for Spain and Mexico, Google is presented with exactly the same content, letter for letter.
The canonical acknowledges that this is the same content. The Hreflang tag identifies which URL should be displayed in different sets of results.
So, in other words, canonical + Hreflang = same content + different URL.
Google knows the content is the same, but displays the correct URL for the Google domain search (e.g. google.com.mx will see the relevant URLs for Mexico displayed in the results).
-
With canonical tag it is a one way road:
You have Page A and Page B with the same content but you want to point out Page A
Page B has a canonical to Page A:
Page B will disappear from the Search Results transferring all the link juice that it has gained to Page A
If you have the same content in different languages then you should use hreflang telling search engines that the two are the same but in other language:
Page A and Page B will have both the following in their headers
This way you will not Geo-Target but Language-Target the two pages ;-)
-
thanks Istvan
but what about whether its a requirement, or suggested best practice, that if you have tags (say canonical) on one set of duplicate pages then you must also add to the other similar/dupe pages (on original site).
Can you have one but not the other without it causing issues or do you need both to stop duplicate issues ?
-
Sorry for responding late, but I somehow forgot to answer this one.
So basically I would consider putting HREFLANG to all of the pages (US, original and any other language). Please note that HREFLANG is connected to optimizing the same content on different languages and not for geo-targeting mainly.
The best example would be Belgium. You can have content in French and in Dutch, still you are optimizing for the same region.
-
Thanks Itsvan, its a good answer and further information! What im really trying to establish though is if its ok to ONLY add canonical & href tags to the US focused subdirectory site ? Do they need to be added to the main site too or can I leave them off (since dont want to geotarget the main site) ? Im confused by wording on google articles/bogs etc on this subject. Since think they say that if you put the tags on a duplicate page you need to also put tags with alternative region/lang tags on the corresponding dupe content page (although i dont want to since want to leave main site free of specific geotargeting). In other words is it a technical requirement/necessity to have tags on both sets of dupe content ?
-
Hi danarchism,
This is what we have on a quite big website:
1. Main site is geo-targeted for a specific country
2. sub-folders of the site are geo-targeted for other countries
3. On each Page in the header we have the HREFLANG to the other 9 languages we use on the site.
Still when we talk about SERP impressions we have many times overlays (Such as the Geo-Targeted content to the Netherlands will appear in the Google.be or Geo-Targeted content to Germany appears in Google.At).
I hope this helped,
Istvan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate content
I have one client with two domains, identical products to appear on both domains. How should I handle this?
Technical SEO | | Hazel_Key0 -
Duplicate Content Issue
My issue with duplicate content is this. There are two versions of my website showing up http://www.example.com/ http://example.com/ What are the best practices for fixing this? Thanks!
Technical SEO | | OOMDODigital0 -
SEOMOZ and non-duplicate duplicate content
Hi all, Looking through the lovely SEOMOZ report, by far its biggest complaint is that of perceived duplicate content. Its hard to avoid given the nature of eCommerce sites that oestensibly list products in a consistent framework. Most advice about duplicate content is about canonicalisation, but thats not really relevant when you have two different products being perceived as the same. Thing is, I might have ignored it but google ignores about 40% of our site map for I suspect the same reason. Basically I dont want us to appear "Spammy". Actually we do go to a lot of time to photograph and put a little flavour text for each product (in progress). I guess my question is, that given over 700 products, why 300ish of them would be considered duplicates and the remaning not? Here is a URL and one of its "duplicates" according to the SEOMOZ report: http://www.1010direct.com/DGV-DD1165-970-53/details.aspx
Technical SEO | | fretts
http://www.1010direct.com/TDV-019-GOLD-50/details.aspx Thanks for any help people0 -
Duplicate Page Content for sorted archives?
Experienced backend dev, but SEO newbie here 🙂 When SEOmoz crawls my site, I get notified of DPC errors on some list/archive sorted pages (appending ?sort=X to the url). The pages all have rel=canonical to the archive home. Some of the pages are shorter (have only one or two entries). Is there a way to resolve this error? Perhaps add rel=nofollow to the sorting menu? Or perhaps find a method that utilizes a non-link navigation method to sort / switch sorted pages? No issues with duplicate content are showing up on google webmaster tools. Thanks for your help!
Technical SEO | | jwondrusch0 -
Problem with duplicate content
Hi, My problem is this: SEOmoz tells me I have duplicate content because it is picking up my index page in three different ways: http://www.web-writer-articles.co.uk http://www.web-writer-articles.co.uk/ and http://www.web-writer-articles.co.uk/index.php Can someone give me some advice as to how I can deal with this issue? thank you for your time, louandel15
Technical SEO | | louandel150 -
Bad Duplicate content issue
Hi, for grappa.com I have about 2700 warnings of duplicate page content. My CMS generates long url like: http://www.grappa.com/deu/news.php/categoria=latest_news/idsottocat=5 and http://www.grappa.com/deu/news.php/categoria%3Dlatest_news/idsottocat%3D5 (this is a duplicated content). What's the best solution to fix this problem? Do I have to set up a 301 redirect for all the duplicated pages or insert the rel=canonical or rel=prev,next ? It's complicated becouse it's a multilingual site, and it's my first time dealing with this stuff. Thanks in advance.
Technical SEO | | nico860 -
Noindex duplicate content penalty?
We know that google now gives a penalty to a whole duplicate if it finds content it doesn't like or is duplicate content, but has anyone experienced a penalty from having duplicate content on their site which they have added noindex to? Would google still apply the penalty to the overall quality of the site even though they have been told to basically ignore the duplicate bit. Reason for asking is that I am looking to add a forum to one of my websites and no one likes a new forum. I have a script which can populate it with thousands of questions and answers pulled direct from Yahoo Answers. Obviously the forum wil be 100% duplicate content but I do not want it to rank for anyway anyway so if I noindex the forum pages hopefully it will not damage the rest of the site. In time, as the forum grows, all the duplicate posts will be deleted but it's hard to get people to use an empty forum so need to 'trick' them into thinking the section is very busy.
Technical SEO | | Grumpy_Carl0 -
Duplicate content across multiple domains
I have come across a situation where we have discovered duplicate content between multiple domains. We have access to each domain and have recently within the past 2 weeks added a 301 redirect to redirect each page dynamically to the proper page on the desired domain. My question relates to the removal of these pages. There are thousands of these duplicate pages. I have gone back and looked at a number of these cached pages in google and have found that the cached pages that are roughly 30 days old or older. Will these pages ever get removed from google's index? Will the 301 redirect even be read by google to be redirected to the proper domain and page? If so when will that happen? Are we better off submitting a full site removal request of the sites that carries the duplicate content at this point? These smaller sites do bring traffic on their own but I'd rather not wait 3 months for the content to be removed since my assumption is that this content is competing with the main site. I suppose another option would be to include no cache meta tag for these pages. Any thoughts or comments would be appreciated.
Technical SEO | | jmsobe0