Multiple Countries, Same Language: Receiving Duplicate Page & Content Errors
-
Hello!
I have a site that serves three English-speaking countries, and is using subfolders for each country version:
- United Kingdom: https://site.com/uk/
- Canada: https://site.com/ca/
- United States & other English-speaking countries: https://site.com/en/
The site displayed is dependent on where the user is located, and users can also change the country version by using a drop-down flag navigation element in the navigation bar. If a user switches versions using the flag, the first URL of the new language version includes a language parameter in the URL, like:
In the Moz crawl diagnostics report, this site is getting dinged for lots of duplicate content because the crawler is finding both versions of each country's site, with and without the language parameter.
However, the site has rel="canonical" tags set up on both URL versions and none of the URLs containing the "?language=" parameter are getting indexed.
So...my questions:
1. Are the Duplicate Title and Content errors found by the Moz crawl diagnostic really an issue?
2. If they are, how can I best clean this up?
Additional notes: the site currently has no sitemaps (XML or HTML), and is not yet using the hreflang tag. I intend to create sitemaps for each country version, like:
- .com/en/sitemap.xml
- .com/ca/sitemap.xml
- .com/uk/sitemap.xml
I thought about putting a 'nofollow' tag on the flag navigation element, but since no sitemaps are in place I didn't want to accidentally cut off crawler access to alternate versions.
Thanks for your help!
-
Yep, given your resource constraints, I'd focus on translations for now. If you ever get to a point that there is something bigger than price differentiating your content, then you can think about geo-targeting. You will need the resources to differentiate the content though.
Right now, my recommendation is to drop the country specific content and just offer English for now. Your content can rank for any English speaking search, regardless of country. However, if the terms people use in the US, UK and Canada differ that much, you can "translate" the content (en-us, en-gb, en-au) and use the HREFLANG tag.
For price changes, that's tricker, but do you offer the price in search results via schema? Does it show up? If not, then you can use cookies to set the prices dependent on the country the person chooses (try not to use IP address, and if you do, make people confirm the setting).
For now, focus your time and efforts getting the flow right for the user. Only worry about HREFLANG if your English content needs to be differentiated for term usage. Then focus your efforts on getting those upcoming translations right. When that is ready, then really use HREFLANG.
Hope that helps!
-
Hi Kate,
Nifty quiz and flowchart! Thanks for sharing it. All the countries targeted are English-speaking, though further expansion to non-English speaking countries is planned for 2015. Here are the answers to the questions:
1. Does your business/product/content change in different countries?
A: Not really. 90% of the products are available in all three countries, and only one country is currently lacking the remaining 10%, and it will start selling those products there in 2015.
2. Would it make sense to an international visitor to see different site content? (ex. currency, localization, etc.)
A: Currency - yes. Otherwise, not really.
3. Do you have the resources to differentiate the content?
A: Not currently. This is a set of branded products, and the product descriptions use extensive "on-brand" language.
4. Are there multiple official languages for any of these countries?
A: Yes, Canada's official languages are English and French. There is no French version currently available.
5. Do you plan on offering the site content in all official languages?
A: Next six months - no. Late 2015 - maybe.
Going through the quiz, if I answer:
1. No, 2. Yes, 3. No
This is the recommendation:
Your International Strategy is:
Translate Only
- Don’t machine translate, while manual translation is costly, it’s the best for your goals.
- Put your HREFLANG in XML sitemaps.
- Use the Language Meta tag for Bing translation targeting.
- Don’t use a ccTLD. That is for Geo-Targeting only.
Aside from the manual translation portion, do you think #2 and #3 are still the best solutions for this situation?
Thanks for your help!
-
Hi!
This is a tough one because I can't tell if you mean to be geo-targeting or translate. It's not a one or the other thing, but it usually is when you are just targeting english speaking countries. Can you do me a favor and go to http://www.katemorris.com/issg/ and go through the questions there? Let me know what the "answer" is for your situation and I'll help you get to the right solution.
But in short, yes, the duplicate content is a real issue with or without the lang parameter.
Let me know!
-
Oh this is a tough one. The problem is that no matter the tags and language, the content is the same. It is reflecting duplicate content because it is duplicate content. Duplicate content within your site is serious, especially if you are trying to target keywords on those pages.
The hreflang tags should help you be able to display languages without using so many duplicate pages. I don't have much experience with that tag, but my advice would be to look into it further to help with your duplicate content issue. No following the duplicate pages will ultimately effect their rankings, so that probably isn't the best thing to do.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What's more valuable: new content or optimizing old content
We are a niche legacy print publication that's been around for close to 20 years. Recently, we combined several old sites in one new responsive site. We have over 7,000 articles -many of which are evergreen and can be repurposed when needed. Most of the old pieces although published, have not been optimized for SEO. However, as we create new pieces, we optimize them for search and social and they tend to get more organic traffic. Where we're torn is on how much we should balance our limited editorial resources between cleaning up and optimizing our extensive archive to improve our organic reach, vs. pumping out new original pieces each week. I realize that without a lot of data the answers will be varied - I guess I'm looking for a best practices approach for content publishers. If it helps at all, our main conversion goal is selling subscriptions to our print and digital publications. We know that organic traffic tends to be more engaged than our social referrals. Unfortunately, due to the nature of the magazine fulfilment business, it's tough to know which channels convert better. Thanks!
Moz Pro | | RicardoSalcedo0 -
What's the best way to eliminate "429 : Received HTTP status 429" errors?
My company website is built on WordPress. It receives very few crawl errors, but it do regularly receive a few (typically 1-2 per crawl) "429 : Received HTTP status 429" errors through Moz. Based on my research, my understand is that my server is essentially telling Moz to cool it with the requests. That means it could be doing the same for search engines' bots and even visitors, right? This creates two questions for me, which I would greatly appreciate your help with: Are "429 : Received HTTP status 429" errors harmful for my SEO? I imagine the answer is "yes" because Moz flags them as high priority issues in my crawl report. What can I do to eliminate "429 : Received HTTP status 429" errors? Any insight you can offer is greatly appreciated! Thanks,
Moz Pro | | ryanjcormier
Ryan0 -
Clearing our on-page ranking reports?
Is there a way to "bulk delete" on-page ranking reports which are no longer relevant? I know we can delete them one at a time, but the reason I ask is that I've done a fair bit of work changing URL's, so the reports are often for old URL's which no longer exist. (yes, I made sure to do 301 redirects to the new ones!) Thanks in advance for any help!
Moz Pro | | koalatm0 -
Duplicate page titles in SEOMoz
My on page reports are showing a good number of duplicate title tags, but they are all because of a url tracking parameter that tells us which link the visitor clicked on. For example, http://www.example.com/example-product.htm?ref=navside and http://www.example.com/example-product.htm are the same page, but are treated as to different urls in SEOMoz. This is creating "fake" number of duplicate page titles in my reports. This has not been a problem with Google, but SEOMoz is treating it like this and it's confusing my data. Is there a way to specify this as a url parameter in the Moz software? Or does anybody have another suggestion? Should I specify this in GWT and BWT?
Moz Pro | | InetAll0 -
SEOmoz duplicate content checker
From my reports in seomoz i can see pages that are showing as having duplicate content but when i click on them it does not show me which pages are carrying the duplicate content? Is there any way to check this via semoz reports?
Moz Pro | | jazavide0 -
Duplicate page content on / and index.php
Hi I am new to SEOmoz and in the crawl diagnostics for one of my clients it came back duplicate content on the homepage www.myclient.co.uk and on the www.myclient.co.uk/index.php which is obviously the same page. I understand that the key is to do a 301 redirect from the index to /, however how will I know that this will not just create an ever ending loop on the server? From your experience how is the best way to tackle this crawl error? Also is there a specific question that I need to ask the server?
Moz Pro | | search_shop0 -
Roger keeps telling me my canonical pages are duplicates
I've got a site that's brand spanking new that I'm trying to get the error count down to zero on, and I'm basically there except for this odd problem. Roger got into the site like a naughty puppy a bit too early, before I'd put the canonical tags in, so there were a couple thousand 'duplicate content' errors. I put canonicals in (programmatically, so they appear on every page) and waited a week and sure enough 99% of them went away. However, there's about 50 that are still lingering, and I'm not sure why they're being detected as such. It's an ecommerce site, and the duplicates are being detected on the product page, but why these 50? (there's hundreds of other products that aren't being detected). The URLs that are 'duplicates' look like this according to the crawl report: http://www.site.com/Product-1.aspx http://www.site.com/product-1.aspx And so on. Canonicals are in place, and have been for weeks, and as I said there's hundreds of other pages just like this not having this problem, so I'm finding it odd that these ones won't go away. All I can think of is that Roger is somehow caching stuff from previous crawls? According to the crawl report these duplicates were discovered '1 day ago' but that simply doesn't make sense. It's not a matter of messing up one or two pages on my part either; we made this site to be dynamically generated, and all of the SEO stuff (canonical, etc.) is applied to every single page regardless of what's on it. If anyone can give some insight I'd appreciate it!
Moz Pro | | icecarats0 -
On Page missing keywords
I setup my keywords on SEOMoz properly but the On Page result just shows me 2 keywords instead of the 7 that I set for my campaign. I was expecting the application to score the other keywords on wednesday but it did not add the missing keywords. Is this a bug?
Moz Pro | | netbuilder0