International Sites and Duplicate Content
-
Hello,
I am working on a project where have some doubts regarding the structure of international sites and multi languages.Website is in the fashion industry. I think is a common problem for this industry. Website is translated in 5 languages and sell in 21 countries.
As you can imagine this create a huge number of urls, so much that with ScreamingFrog I cant even complete the crawling.
Perhaps the UK site is visible in all those versions
http://www.MyDomain.com/en/GB/
http://www.MyDomain.com/it/GB/
http://www.MyDomain.com/fr/GB/
http://www.MyDomain.com/de/GB/
http://www.MyDomain.com/es/GB/
Obviously for SEO only the first version is important
One other example, the French site is available in 5 languages and again...
http://www.MyDomain.com/fr/FR/
http://www.MyDomain.com/en/FR/
http://www.MyDomain.com/it/FR/
http://www.MyDomain.com/de/FR/
http://www.MyDomain.com/es/FR/
And so on...this is creating 3 issues mainly:
-
Endless crawling - with crawlers not focusing on most important pages
-
Duplication of content
-
Wrong GEO urls ranking in Google
I have already implemented href lang but didn't noticed any improvements. Therefore my question is
Should I exclude with "robots.txt" and "no index" the non appropriate targeting?
Perhaps for UK leave crawable just English version i.e. http://www.MyDomain.com/en/GB/, for France just the French version http://www.MyDomain.com/fr/FR/ and so on
What I would like to get doing this is to have the crawlers more focused on the important SEO pages, avoid content duplication and wrong urls rankings on local Google
Please comment
-
-
Hey Guido, don't know if it's the best solution, but could be a temporary fix until the best solution is in place. I suggest to move forward with proper HREF LANG tagging or definitely delete those irrelevant languages. Try to do what I said before about validate each country/language and submit a sitemap.xml reflecting that folder to see crawl and index stats pero country/language. Add a sitemap index and obviously validate your entire domain. Just block in the robots.txt unnecessary folders, like images, js libraries, etc. to save crawl budget to your domain.
Let me know if you have another doubt
-
Thank you Antonio, insightful and clear.
There is really not a need of EN versions of localized sites, I think has been done more as was easier to implement (original site is EN-US).
Don't you think robots and noindex EN version of localized sites could be the best solution? for sure is the easier one to implement without affecting UX.
-
Don't know why you have a UK oriented site for German and Italian people, I think is not important those languages in a country mainly English speaking (not US for example, there you must have a Spanish version, or in Canada for English and French). The owner must have their reasons.
Besides this, about your questions:
- If those non-relevant languages must live there, it's correct to implement HREF LANG (may take some time to show results). Also, if the domain is gTLD, you can validate all the subfolders in Google Search Console and choose the proper International targeting. With the ammount of languages and countries I imagine this might be a pain in the ***.
- About the crawling, for large sitesI recommend to crawl per language. If neccesary, per language-country. In this instance I recommend to create a sitemap XML per language or language-country for just HTML pages (hopefully dynamically updated by the e-commerce), create a Sitemap Index in the root of the domain and submit them in Google Search Console (better if you validated the languages or language-country). With this you can answer the question if some language or country are being not indexed with the Submited/Indexed stadistics of GSC.
- Maybe the robots.txt might save your crawl budget, but I'm not a fan of de-index if those folders are truly not relevant (after all, there should be a italian living in UK. If you can't delete the irrelevant langauges for some countries, this can be an option
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is it worth maintaining multiple international websites
Hi I work for a British company which has two well established websites - a .co.Uk for the UK, and a .com for the US and rest of the world (in language directories). The Uk site is hosted in the Uk, the .com in US. The websites do reasonable well in Google on both sides of the Atlantic. The company is a small but quite well known brand. The company is now thinking of redirecting the .co.Uk to the .com as it would be cheaper to maintain. What would you advise? Thanks.
International SEO | | fdl4712_aol.com2 -
Is using JavaScript to render translations safe for International SEO?
Hello World! Background: I am evaluating a tool/service that a company wants to use for managing the translated versions of their international/multi-lingual websites: https://www.transifex.com/product/transifexlive/ Transifex is asking webmaster to "simply add a snippet of JavaScript" to their website(s); the approved translations are added by the business in the back-end; and the translated sites are made live with the click of a button (on/to the proper ccTLD, sub-domain, or sub-directory, which is specified). CONCERN: Even though I know Google reads JavaScript for crawling and ranking,
International SEO | | SixSpokeMedia64
I am concerned because I see the "English text" when I view the source-code on the "German site", and I wonder if this is really acceptable? QUESTION: Is a service like this (such as Transifex using JavaScript to render translations client-side) safe for indexing and ranking for my clients' international search engine visibility, especially via Google? Thank you!0 -
Rel alternate use for multi regional website with english language and same content
I have a website targeting multi language and multi regional users. For example, my main site is www.abcd.com which is in English and targeting US. I have the same content in English which is targeting UK and India with www.uk.abcd.com and www.india.abcd.com. I want to avoid content duplication and help search engines to show the right pages on the country level searches. I have researched a bit and have come to conclusion of using re alternate tag. Can someone help me with how to place the codes for the same. Many thanks Mozers!
International SEO | | HiteshBharucha0 -
B2B International Subdirectory - How Unique is Unique?
With the power to upload unique xml sitemaps for a subdirectory targeting the UK, geo-target the UK in WebmasterTools, and the ability to adjust content to adhere to en-gb standards, would it still be essential for a site to re-write all of it's content if it wanted to rank well or could we just use the same content as our en-us pages with the dialect changes and other tools mentioned above? Not interested in unique TLDs or subdomains.
International SEO | | SEOPPCDP0 -
Keyphrase ranking a geo-redirected site in Google
Hi all This is the situation. I have a client who runs a number of ccTLD sites (all exact match brand name domains), including a .com which they use for the US. This is a hair care product and due to Advertising Standards Authority (UK) restrictions, they cannot use a certain phrase to promote their products - 'hair loss' on the domain.co.uk site. However, in the US, there is no such restriction and can use wording this on the site. A brand name search in google.co.uk brings up .co.uk as 1st result and .com as 2nd result, so the .com is indexed in google.co.uk. Any non-US user visiting domain.com will be redirected to their ccTLD site. Here's my question - could I feasibly get the domain.com site ranking in google.co.uk for certain 'hair loss' based keyphrases, considering the fact that I can mention it in the copy on there but not on the domain.co.uk site. Would I need to remove any Geographic Target in the WMT account for domain.com? Or is this a form of Google cloaking and could see the site penalised? Thanks
International SEO | | Coolpink0 -
International Link Building - France, Spain, Germany, Italy, Switzerland
I've got a partner agency (non-SEO) in Europe who wants to send some additional SEO business our way, but I don't currently have a system in place geared specifically towards international, country specific link building. Does anyone know of any resources (blogs, lists, tools) specifically geared towards getting links from country specific TLDs for France, Spain, Germany, Italy and Switzerland? (.fr, .es, .de, .it and .ch are the TLDs.) .co.uk sources would also be handy. A list of potential link building sources in those countries would be most helpful. I fully understand the SEO elements in play for international SEO, I just don't have any decent resource lists for those specific countries. Sites in those countries that accept guest blog posts, language specific infographic sites, foreign PR platforms, high-quality non-penalized directories...really anything would be awesome! Thanks in advance folks!
International SEO | | Point_It0 -
Am I doing this right? Same website, content and similar domains.
I have 5 sites with the same exact content. I have a separate webmastertool for each one and I have targeted them to each country on WebMastertools? Iam I missing something or did I do it right.Thankswww.abc.com (USA)www.abc.com.ar (argentina)www.abc.com.mx (mexico)www.abc.com.co (colombia)
International SEO | | M_80 -
What is the best SEO site structure for multi country targeting?
Hi There, We are an online retailer with four (and soon to be five) distinct geographic target markets (we have physical operations in both the UK and New Zealand). We currently target these markets like this: United Kingdom (www.natureshop.co.uk) New Zealand (www.natureshop.co.nz) Australia (www.natureshop.com/au) - using a google web master tools geo targeted folder United States (www.natureshop.com) - using google web master tools geo targeted domain Germany (www.natureshop.de) - in german and yet to be launched as full site We have various issues we want to address. The key one is this: our www.natureshop.co.uk website was adversely affected by the panda update on April 12. We had some external seo firms work on this site for us and unfortunately the links they gained for us were very low quality, from sometimes spammy sites and also "keyword" packed with very littlle anchor text variation. Our other websites (the .co.nz and .com) moved up after the updates so I can only assume our external seo consultants were responsible for this. I have since managed to get them to remove around 70% of these links and we have bought all seo efforts back in house again. I have also worked to improve the quality of our content on this site and I have 404'ed the six worst affected pages (the ones that had far too many single phrase anchor text links coming into them). We have however not budged much in our rankings (we have made some small gains but not a lot). Our other weakness's are not the fastest page load times and some "thin" content. We are on the cusp (around 4 weeks away) of deploying a brand new platform using asp.net MVP with N2 and this looks like it will address our page load speed issues. We also have been working hard on our content building and I believe we will address that as well with this release. Sorry for the long build up, however I felt some background was needed to get to my questions. My questions are: Do you think we are best to proceed with trying to get our www.natureshop.co.uk website out of the panda trap or should we consider deploying a new version of the site on www.natureshop.com/uk/ (geo targeted to the UK)? If we are to do this should we do the same for New Zealand and Germany and redirect the existing domains to the new geo targeted folders? If we do this should we redirect the natureshop.co.uk pages to the new www.natureshop.com/uk/ pages or will this simply pass on the panda "penalty". Will this model build stronger authority on the .com domain that benefit all of the geo targeted sub folders or does it not work this way? Finally can we deploy the same pages and content on the different geo targeted sub folders (with some subtle regional variations of spelling and language) or will this result in a duplicate content penalty? Thank you very much in advance to all of you and I apologise for the length and complexity of the question. Kind Regards
International SEO | | ConradC
Conrad Cranfield
Founder: Nature Shop Ltd0