International web site - duplicate content?
-
I am looking at a site offering different language options via a javascript drop down chooser. Will google flag this as duplicate content? Should I recommend the purchase of individual domains for each country?
i.e. .uk
-
To avoid duplicate content you need to use the rel=”alternate” href=”x” tag. You dont have to buy a domain, here are several ways you can organise your website.
ccTLDs - [example.ie]
Pros
- Clear geotargeting
- Server location irrelevant
- Easy separation of sites
Cons
- Expensive (and may have limited availability)
- Requires more infrastructure
- Strict ccTLD requirements (sometimes)
****Subdomains with gTLDS [de.example.com]
Pros
- Easy to set up
- Can use Webmaster Tools geotargeting
- Allows different server locations
- Easy separation of sites
Cons
- Users might not recognize geotargeting from the URL alone (is “de” the language or country?)
Subdirectories with gTLDs [example.com/de/]
Pros
- Easy to set up
- Can use Webmaster Tools geotargeting
- Low maintenance (same host)
Cons
- Users might not recognize geotargeting from the URL alone
- Single server location
- Separation of sites harder
URL parameters [site.com?loc=de]
Pros
- Not recommended.
Cons
- URL-based segmentation difficult
- Users might not recognize geotargeting from the URL alone
- Geotargeting in Webmaster Tools is not possible
You can have even the same content for example, if you have a version for the UK and another for the US the content will be very similar.
Hope this helps
-
It's been my experience that if you're going to have additional languages on your site, you can do one of a few things to the URL:
or
or
Then you can place the content in the respective languages on those sites. They won't necessarily be seen as duplicated content (semantics change as you translate to different languages). However, Google won't necessarily penalize you if you keep your URL to site.com and just switch up the languages because, again, semantics change as you translate languages.
I worked with several sites that translated content from English to Spanish or English to Hebrew and they were never once penalized for duplicated content.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What to do with internal spam url's google indexed?
I am in SEO for years but never met this problem. I have client who's web page was hacked and there was posted many, hundreds of links, These links has been indexed by google. Actually these links are not in comments but normal external urls's. See picture. What is the best way to remove them? use google disavow tool or just redirect them to some page? The web page is new, but ranks good on google and has domain authority 24. I think that these spam url's improved rankings too 🙂 What would be the best strategy to solve this. Thanks. k9Bviox
White Hat / Black Hat SEO | | AndrisZigurs0 -
Schema Markup for regular web pages?
I'm a bit confused about what Schema markup should be applied to such regular, informative web pages.
White Hat / Black Hat SEO | | gray_jedi
We have a few pages describing our technology and solutions. These pages are not products or news articles. And they are not something that should be reviewed/rated. What Schema markup should be used for a standard run-of-the mill web page?
Is there a good reference / tutorial for optimizing the schema markup of an informational website? Any advice is much appreciated, thank you!0 -
Why is this site performing so well in the SERP's and getting high traffic volume for no apparent reason!
The site is https://virtualaccountant.ie/ It's a really small site They have only about 7 back links, They don't blog They don't have a PPC campaign They don't stand out from the crowd in terms of product or services offered So why are they succeeding in topping the SERP's for difficult to rank for accounting keywords such as accountant and online accounts. What are they doing better than everyone else, or have they discovered a way to cheat Google, and worse still - ME!
White Hat / Black Hat SEO | | PeterConnor0 -
Hiding content or links in responsive design
Hi, I found a lot of information about responsive design and SEO, mostly theories no real experiment and I'd like to find a clear answer if someone tested that. Google says:
White Hat / Black Hat SEO | | NurunMTL
Sites that use responsive web design, i.e. sites that serve all devices on the same set of URLs, with each URL serving the same HTML to all devices and using just CSS to change how the page is rendered on the device
https://developers.google.com/webmasters/smartphone-sites/details For usability reasons sometimes you need to hide content or links completely (not accessible at all by the visitor) on your page for small resolutions (mobile) using CSS ("visibility:hidden" or "display:none") Is this counted as hidden content and could penalize your site or not? What do you guys do when you create responsive design websites? Thanks! GaB0 -
Multiple domains different content same keywords
what would you advice on my case: It is bad for google if i have the four domains. I dont link between them as i dont want no association, or loss in rakings in branded page. Is bad if i link between them or the non branded to them branded domain. Is bad if i have all on my webmaster tools, i just have the branded My google page is all about the new non penalized domain. altough google gave a unique domain +propdental to the one that he manually penalized. (doesn't make sense) So. What are the thinks that i should not do with my domain to follow and respect google guidelines. As i want a white hat and do not do something that is wrong without knowledge
White Hat / Black Hat SEO | | maestrosonrisas0 -
Duplicate Content
Hi, I have a website with over 500 pages. The website is a home service website that services clients in different areas of the UK. My question is, am I able to take down the pages from my URL, leave them down for say a week, so when Google bots crawl the pages, they do not exist. Can I then re upload them to a different website URL, and then Google wont penalise me for duplicate content? I know I would of lost juice and page rank, but that doesnt really matter, because the site had taken a knock since the Google update. Thanks for your help. Chris,
White Hat / Black Hat SEO | | chrisellett0 -
What on-page/site optimization techniques can I utilize to improve this site (http://www.paradisus.com/)?
I use a Search Engine Spider Simulator to analyze the homepage and I think my client is using black hat tactics such as cloaking. Am I right? Any recommendations on to improve the top navigation under Resorts pull down. Each of the 6 resorts listed are all part of the Paradisus brand, but each resort has their own sub domain.
White Hat / Black Hat SEO | | Melia0 -
Somebody hacked many sites and put links to my sites in hidden div
I had 300 good natural links to my site from different sites and site ranked great for my keywords. Somebody (I suppose my competitor) has hacked other sites 2 days ago (checked Google cache) and now Yahoo Site Explorer shows 600 backlinks. I've checked new links - they all are in the same hidden div block - top:-100px; position:absolute;. I'm afraid that Google may penalize my site for these links. I'm contacting webmasters of these sites and their hosting so they remove these links. Is it possible to give Google a notice that these links are not mine so it could just skip them not penalizing me? Is it safe to make "Spam report" regarding links to my own site?
White Hat / Black Hat SEO | | zarades0