Geotargeting duplicate content to different regions - href and canonical tag confusion
-
If you duplicate content onto a sub-folder for say a new US geotargeted site (to target kw spelling differences) and, in addition to GWT geotargeting settings, implement the 'Canonical' and 'Hreflang' tags on these new pages to show G different region and language version (en-us). Then does the original/main site similar pages also need to have canonical and href tags ?
The main/original sites page I don't really want to target a specific country (although existing signals (hosting etc) will be UK (primary target of main site) but pages show up in other country searches too (which we want).
Im presuming fine to leave the original/main site as it currently is although wording in google blog/webmaster central articles etc are a bit confusing hence why im asking for anyone elses opinion/input on this.
Also is there are any benefit (or just best practice) to use 'www.example.com/en-us/...' in the subdirectory URL as opposed to just 'www.example.com/us/'
many thanks in advance to any commentators
-
Many thanks Gianluca !!
-
Hi,
I suggest you both to give a read to this post by DejanSEO, which is quite clear and - IMHO - points to the right interpretation of a somehow confused best practice.
-
Thats what i thought originally but getting confised when i read this page: http://googlewebmastercentral.blogspot.com.au/2011/12/new-markup-for-multilingual-content.html
Specifically this bit:
Annotating pages as substantially similar content
Optionally, for pages that have substantially the same content in the same language and are targeted at multiple countries, you may use the rel="canonical" link element to specify your preferred version. We’ll use that signal to focus on that version in search, while showing the local URLs to users where appropriate. For example, you could use this if you have the same product page in German, but want to target it separately to users searching on the Google properties for Germany, Austria, and Switzerland.
And read in conjunction with this article:
Specifically this bit:
The Effect Of Combining Canonical Tags & Hreflang Tags
Not forgetting that the canonical tags should only be used with content in the same language, when would we use both?
Well firstly, the use of both would involve what I usually call world languages such as English, Spanish, French or Portuguese. These languages are used in many countries and, whilst there are variations between the use of these languages in those countries, the variations are sometimes small.
Additionally, multinational publishers often save costs by using one version of the language for all countries speaking that general language, thus ignoring the regional variations. In other words, for Spain and Mexico, Google is presented with exactly the same content, letter for letter.
The canonical acknowledges that this is the same content. The Hreflang tag identifies which URL should be displayed in different sets of results.
So, in other words, canonical + Hreflang = same content + different URL.
Google knows the content is the same, but displays the correct URL for the Google domain search (e.g. google.com.mx will see the relevant URLs for Mexico displayed in the results).
-
With canonical tag it is a one way road:
You have Page A and Page B with the same content but you want to point out Page A
Page B has a canonical to Page A:
Page B will disappear from the Search Results transferring all the link juice that it has gained to Page A
If you have the same content in different languages then you should use hreflang telling search engines that the two are the same but in other language:
Page A and Page B will have both the following in their headers
This way you will not Geo-Target but Language-Target the two pages ;-)
-
thanks Istvan
but what about whether its a requirement, or suggested best practice, that if you have tags (say canonical) on one set of duplicate pages then you must also add to the other similar/dupe pages (on original site).
Can you have one but not the other without it causing issues or do you need both to stop duplicate issues ?
-
Sorry for responding late, but I somehow forgot to answer this one.
So basically I would consider putting HREFLANG to all of the pages (US, original and any other language). Please note that HREFLANG is connected to optimizing the same content on different languages and not for geo-targeting mainly.
The best example would be Belgium. You can have content in French and in Dutch, still you are optimizing for the same region.
-
Thanks Itsvan, its a good answer and further information! What im really trying to establish though is if its ok to ONLY add canonical & href tags to the US focused subdirectory site ? Do they need to be added to the main site too or can I leave them off (since dont want to geotarget the main site) ? Im confused by wording on google articles/bogs etc on this subject. Since think they say that if you put the tags on a duplicate page you need to also put tags with alternative region/lang tags on the corresponding dupe content page (although i dont want to since want to leave main site free of specific geotargeting). In other words is it a technical requirement/necessity to have tags on both sets of dupe content ?
-
Hi danarchism,
This is what we have on a quite big website:
1. Main site is geo-targeted for a specific country
2. sub-folders of the site are geo-targeted for other countries
3. On each Page in the header we have the HREFLANG to the other 9 languages we use on the site.
Still when we talk about SERP impressions we have many times overlays (Such as the Geo-Targeted content to the Netherlands will appear in the Google.be or Geo-Targeted content to Germany appears in Google.At).
I hope this helped,
Istvan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate Content Issues - Where to start???
Dear All I have recently joined a new company Just Go Holidays - www.justgoholidays.com I have used the SEO Moz tools (yesterday) to review the site and see that I have lots of duplicate content/pages and also lots of duplicate titles all of which I am looking to deal with. Lots of the duplicate pages appear to be surrounding, additional parameters that are used on our site to refine and or track various marketing campaigns. I have therefore been into Google Webmaster Tools and defined each of these parameters. I have also built a new XML sitemap and submitted that too. It looks as is we have two versions of the site, one being at www.justgoholidays.com and the other without the www It appears that there are no redirects from the latter to the former, do I need to use 301's here or is it ok to use canonicalisation instead? Any thoughts on an action plan to try to address these issues in the right order and the right way would be very gratefully received as I am feeling a little overwhelmed at the moment. (we also use a CMS system that is not particularly friendly and I think I will have to go directly to the developers to make lots of the required changes which is sure to cost - therefore really don't want to get this wrong) All the best Matt
Technical SEO | | MattByrne0 -
How different should content be so that it is not considered duplicate?
I am making a 2nd website for the same company. The name of the company, our services, keywords and contact info will show up several times within the text of both websites. The overall text and paragraphs will be different but some info may be repeated on both sites. Should I continue this? What precautions should I take?
Technical SEO | | savva0 -
Duplicate pages in Google index despite canonical tag and URL Parameter in GWMT
Good morning Moz... This is a weird one. It seems to be a "bug" with Google, honest... We migrated our site www.three-clearance.co.uk to a Drupal platform over the new year. The old site used URL-based tracking for heat map purposes, so for instance www.three-clearance.co.uk/apple-phones.html ..could be reached via www.three-clearance.co.uk/apple-phones.html?ref=menu or www.three-clearance.co.uk/apple-phones.html?ref=sidebar and so on. GWMT was told of the ref parameter and the canonical meta tag used to indicate our preference. As expected we encountered no duplicate content issues and everything was good. This is the chain of events: Site migrated to new platform following best practice, as far as I can attest to. Only known issue was that the verification for both google analytics (meta tag) and GWMT (HTML file) didn't transfer as expected so between relaunch on the 22nd Dec and the fix on 2nd Jan we have no GA data, and presumably there was a period where GWMT became unverified. URL structure and URIs were maintained 100% (which may be a problem, now) Yesterday I discovered 200-ish 'duplicate meta titles' and 'duplicate meta descriptions' in GWMT. Uh oh, thought I. Expand the report out and the duplicates are in fact ?ref= versions of the same root URL. Double uh oh, thought I. Run, not walk, to google and do some Fu: http://is.gd/yJ3U24 (9 versions of the same page, in the index, the only variation being the ?ref= URI) Checked BING and it has indexed each root URL once, as it should. Situation now: Site no longer uses ?ref= parameter, although of course there still exists some external backlinks that use it. This was intentional and happened when we migrated. I 'reset' the URL parameter in GWMT yesterday, given that there's no "delete" option. The "URLs monitored" count went from 900 to 0, but today is at over 1,000 (another wtf moment) I also resubmitted the XML sitemap and fetched 5 'hub' pages as Google, including the homepage and HTML site-map page. The ?ref= URls in the index have the disadvantage of actually working, given that we transferred the URL structure and of course the webserver just ignores the nonsense arguments and serves the page. So I assume Google assumes the pages still exist, and won't drop them from the index but will instead apply a dupe content penalty. Or maybe call us a spam farm. Who knows. Options that occurred to me (other than maybe making our canonical tags bold or locating a Google bug submission form 😄 ) include A) robots.txt-ing .?ref=. but to me this says "you can't see these pages", not "these pages don't exist", so isn't correct B) Hand-removing the URLs from the index through a page removal request per indexed URL C) Apply 301 to each indexed URL (hello BING dirty sitemap penalty) D) Post on SEOMoz because I genuinely can't understand this. Even if the gap in verification caused GWMT to forget that we had set ?ref= as a URL parameter, the parameter was no longer in use because the verification only went missing when we relaunched the site without this tracking. Google is seemingly 100% ignoring our canonical tags as well as the GWMT URL setting - I have no idea why and can't think of the best way to correct the situation. Do you? 🙂 Edited To Add: As of this morning the "edit/reset" buttons have disappeared from GWMT URL Parameters page, along with the option to add a new one. There's no messages explaining why and of course the Google help page doesn't mention disappearing buttons (it doesn't even explain what 'reset' does, or why there's no 'remove' option).
Technical SEO | | Tinhat0 -
Avoiding duplicate content on product pages?
Hi, I'm creating a bunch of product pages for courses for a university and I'm concerned about duplicate content penalties. While the page names are different and some of the test is different, much of the text is the same between pairs of pages. I.e. a BA and an MA in a particular subject (say 'hairdressing' will have the same subject descriptions, school introduction paragraph, industry overview paragraph etc. 1. Is this a problem? In a site with 100 pages, if sets of 2 pages have about 50% identical content... 2. If it is a problem, is there anything I can do, other than rewrite the text? 3. From a search perspective, would both pages show up in search results in searches related to 'hairdressing courses' 'study hairdressing' etc? Thanks!
Technical SEO | | AISFM0 -
How to fix duplicate page content error?
SEOmoz's Crawl Diagnostics is complaining about a duplicate page error. The example of links that has duplicate page content error are http://www.equipnet.com/misc-spare-motors-and-pumps_listid_348855 http://www.equipnet.com/misc-spare-motors-and-pumps_listid_348852 These are not duplicate pages. There are some values that are different on both pages like listing # , equipnet tag # , price. I am not sure how do highlight the different things the two page has like the "Equipment Tag # and listing #". Do they resolve if i use some style attribute to highlight such values on page? Please help me with this as i am not really sure why seo is thinking that both pages have same content. Thanks !!!
Technical SEO | | RGEQUIPNET0 -
Duplicate Content based on www.www
In trying to knock down the most common errors on our site, we've noticed we do have an issue with dupicate content; however, most of the duplicate content errors are due to our site being indexed with www.www and not just www. I am perplexed as to how this is happening. Searching through IIS, I see nothing that would be causing this, and we have no hostname records setup that are www.www. Does anyone know of any other things that may cause this and how we can go about remedying it?
Technical SEO | | CredA0 -
Magento Canonical Tags
Magento pages have been giving me a lot of trouble with the canonical tags. In some cases duplicate pages are showing up, so I need to add the canonical tag. In other cases I'm getting an error that there are multiple canonical tags per page. How can I get my pages canonized without duplicate tags? It seems like it's either too much or not enough, no matter what I do. Note: this only applies to category and product pages.
Technical SEO | | GravitateOnline0 -
Duplicate content question with PDF
Hi, I manage a property listing website which was recently revamped, but which has some on-site optimization weaknesses and issues. For each property listing like http://www.selectcaribbean.com/property/147.html there is an equivalent PDF version spidered by google. The page looks like this http://www.selectcaribbean.com/pdf1.php?pid=147 my question is: Can this create a duplicate content penalty? If yes, should I ban these pages from being spidered by google in the robots.txt or should I make these link nofollow?
Technical SEO | | multilang0