Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How do I geo-target continents & avoid duplicate content?
-
Hi everyone,
We have a website which will have content tailored for a few locations:
USA: www.site.com
Europe EN: www.site.com/eu
Canada FR: www.site.com/fr-caLink hreflang and the GWT option are designed for countries. I expect a fair amount of duplicate content; the only differences will be in product selection and prices.
What are my options to tell Google that it should serve www.site.com/eu in Europe instead of www.site.com? We are not targeting a particular country on that continent.
Thanks!
-
Moz most definitively need a "give a beer" feature!! Thanks for the in-depth response. We'll also work on building "local" links as you suggest.
We've since changed the structure of the site to :
USA/Canada: www.site.com
Europe EN: www.site.com/en_gb/
Europe FR: www.site.com/fr_fr/
Canada FR: www.site.com/fr/That way we can use hreflang and avoid duplicate content. In your experience, will Google serve www.site.com/fr_fr/ instead of www.site.com/fr/ to Belgium and Switzerland? Will UK and Ireland see www.site.com or www.site.com/en_gb/ ?
Thanks a lot for the answer!
-
Hi there,
As Marcus mentioned before, at the moment geographical targeting is country based, not per continent, so you're correct: hreflang works for languages or / and countries and the geotarget option in Google Webmaster Tools (when you're not using a ccTLD) is only for countries.
So there are really two alternatives: language targeting (although each language is different in each country) or country targeting (which is the ideal in order to connect with each audience, localizing the content as maximum and leveraging all types of local characteristics).
With language targeting you will avoid having content duplication issues (since it will be only one English or one Spanish version), nonetheless, as I mentioned, it can be tricky: The Spanish spoken in Spain is different than the one from Mexico and each other Latin American country. Seasonality and currency are different. People's culture, tastes and local characteristics too. So language based versions might serve to have a "generic" approach to these audience but not really targeting them as specific markets.
On the other hand with country targeting if you have two English versions you can refer each one to the appropriate country with hreflang, ccTLDs (if you use a generic domain, then with the geotarget option in Google Webmaster tool) and then by doing local link building focused on each country, to enhance the popularity of each version there. This would be the recommended approach. If you can't enable many countries because of resources restrictions then start with the most important ones.
More over, from what you mention about targeting Europe as a whole, even if you enable a domain of the type: www.yourbrand.eu for Europe, it is likely to be treated as a generic domain as Google specifies here, and then inside this domain what you would really have --as I understand from your description-- are language versions targeting Europe in General:
- www.yourbrand.eu/ in English (UK, Ireland, etc.)
- www.yourbrand.eu/fr/ in French (In France, Belgium, Switzerland)
- www.yourbrand.eu/es/ in Spanish
- www.yourbrand.eu/de/ in German (for Germany, Switzerland or Austria)
The issue comes when you have the same content in English for your American audience in www.yourbrand.com or in Spanish (for Spanish speakers in the US) in www.yourbrand.com/es/ that could cause a content duplication issue with www.yourbrand.eu/ and www.yourbrand.eu/es/.
If this is the scenario, then the best you can do is to differentiate the content, changing them by giving signals that one is targeting the US audience and the other, well, what would be English speakers in Europe. But again, there's no real support or straight-forward solution for this scenario since beyond what Google supports, is not "natural" or the best alternative from an "international audience targeting" perspective.
If you have any other information that you think would be relevant to give you additional recommendations please let me know.
I hope this helps!
-
Hey Axial
As far as I am aware there is no option to target regions like Europe and to do this in webmaster tools you will need to create a folder for each country you are looking to target within Europe.
Obviously, there are lots of different languages across Europe so in an ideal world, you will want a version geotargeted to each country in the correct language. If you want to be really fancy you will want a version with english and the relevant countries language.
So, for spain as an example, targeting Spanish and English the hreflang would be set as "ES-es" and "ES-en" (Spain-Spanish and Spain-English). Directories could be matched /es-es & /es-en.
Not an answer as such but as far as I am aware, Europe is not targetable in a single folder via webmaster tools so you are going to have to work with what's available.
Hope that helps
Marcus
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Case Sensitive URLs, Duplicate Content & Link Rel Canonical
I have a site where URLs are case sensitive. In some cases the lowercase URL is being indexed and in others the mixed case URL is being indexed. This is leading to duplicate content issues on the site. The site is using link rel canonical to specify a preferred URL in some cases however there is no consistency whether the URLs are lowercase or mixed case. On some pages the link rel canonical tag points to the lowercase URL, on others it points to the mixed case URL. Ideally I'd like to update all link rel canonical tags and internal links throughout the site to use the lowercase URL however I'm apprehensive! My question is as follows: If I where to specify the lowercase URL across the site in addition to updating internal links to use lowercase URLs, could this have a negative impact where the mixed case URL is the one currently indexed? Hope this makes sense! Dave
Intermediate & Advanced SEO | | allianzireland0 -
Artist Bios on Multiple Pages: Duplicate Content or not?
I am currently working on an eComm site for a company that sells art prints. On each print's page, there is a bio about the artist followed by a couple of paragraphs about the print. My concern is that some artists have hundreds of prints on this site, and the bio is reprinted on every page,which makes sense from a usability standpoint, but I am concerned that it will trigger a duplicate content penalty from Google. Some people are trying to convince me that Google won't penalize for this content, since the intent is not to game the SERPs. However, I'm not confident that this isn't being penalized already, or that it won't be in the near future. Because it is just a section of text that is duplicated, but the rest of the text on each page is original, I can't use the rel=canonical tag. I've thought about putting each artist bio into a graphic, but that is a huge undertaking, and not the most elegant solution. Could I put the bio on a separate page with only the artist's info and then place that data on each print page using an <iframe>and then put a noindex,nofollow in the robots.txt file?</p> <p>Is there a better solution? Is this effort even necessary?</p> <p>Thoughts?</p></iframe>
Intermediate & Advanced SEO | | sbaylor0 -
Duplicate content on subdomains.
Hi Mozer's, I have a site www.xyz.com and also geo targeted sub domains www.uk.xyz.com, www.india.xyz.com and so on. All the sub domains have the content which is same as the content on the main domain that is www.xyz.com. So, I want to know how can i avoid content duplication. Many Thanks!
Intermediate & Advanced SEO | | HiteshBharucha0 -
How to Remove Joomla Canonical and Duplicate Page Content
I've attempted to follow advice from the Q&A section. Currently on the site www.cherrycreekspine.com, I've edited the .htaccess file to help with 301s - all pages redirect to www.cherrycreekspine.com. Secondly, I'd added the canonical statement in the header of the web pages. I have cut the Duplicate Page Content in half ... now I have a remaining 40 pages to fix up. This is my practice site to try and understand what SEOmoz can do for me. I've looked at some of your videos on Youtube ... I feel like I'm scrambling around to the Q&A and the internet to understand this product. I'm reading the beginners guide.... any other resources would be helpful.
Intermediate & Advanced SEO | | deskstudio0 -
News sites & Duplicate content
Hi SEOMoz I would like to know, in your opinion and according to 'industry' best practice, how do you get around duplicate content on a news site if all news sites buy their "news" from a central place in the world? Let me give you some more insight to what I am talking about. My client has a website that is purely focuses on news. Local news in one of the African Countries to be specific. Now, what we noticed the past few months is that the site is not ranking to it's full potential. We investigated, checked our keyword research, our site structure, interlinking, site speed, code to html ratio you name it we checked it. What we did pic up when looking at duplicate content is that the site is flagged by Google as duplicated, BUT so is most of the news sites because they all get their content from the same place. News get sold by big companies in the US (no I'm not from the US so cant say specifically where it is from) and they usually have disclaimers with these content pieces that you can't change the headline and story significantly, so we do have quite a few journalists that rewrites the news stories, they try and keep it as close to the original as possible but they still change it to fit our targeted audience - where my second point comes in. Even though the content has been duplicated, our site is more relevant to what our users are searching for than the bigger news related websites in the world because we do hyper local everything. news, jobs, property etc. All we need to do is get off this duplicate content issue, in general we rewrite the content completely to be unique if a site has duplication problems, but on a media site, im a little bit lost. Because I haven't had something like this before. Would like to hear some thoughts on this. Thanks,
Intermediate & Advanced SEO | | 360eight-SEO
Chris Captivate0 -
Duplicate Content on Wordpress b/c of Pagination
On my recent crawl, there were a great many duplicate content penalties. The site is http://dailyfantasybaseball.org. The issue is: There's only one post per page. Therefore, because of wordpress's (or genesis's) pagination, a page gets created for every post, thereby leaving basically every piece of content i write as a duplicate. I feel like the engines should be smart enough to figure out what's going on, but if not, I will get hammered. What should I do moving forward? Thanks!
Intermediate & Advanced SEO | | Byron_W0 -
Duplicate Content on Press Release?
Hi, We recently held a charity night in store. And had a few local celebs turn up etc... We created a press release to send out to various media outlets, within the press release were hyperlinks to our site and links on certain keywords to specific brands on our site. My question is, should we be sending a different press release to each outlet to stop the duplicate content thing, or is sending the same release out to everyone ok? We will be sending approx 20 of these out, some going online and some not. So far had one local paper website, a massive football website and a local magazine site. All pretty much same content and a few pics. Any help, hints or tips on how to go about this if I am going to be sending out to a load of other sites/blogs? Cheers
Intermediate & Advanced SEO | | YNWA0 -
Concerns about duplicate content issues with australian and us version of website
My company has an ecommerce website that's been online for about 5 years. The url is www.betterbraces.com. We're getting ready to launch an australian version of the website and the url will be www.betterbraces.com.au. The australian website will have the same look as the US website and will contain about 200 of the same products that are featured on the US website. The only major difference between the two websites is the price that is charged for the products. The australian website will be hosted on the same server as the US website. To ensure Australians don't purchase from the US site we are going to have a geo redirect in place that sends anyone with a AU ip address to the australian website. I am concerned that the australian website is going to have duplicate content issues. However, I'm not sure if the fact that the domains are so similar coupled with the redirect will help the search engines understand that these sites are related. I would appreciate any recommendations on how to handle this situation to ensure oue rankings in the search engines aren't penalized. Thanks in advance for your help. Alison French
Intermediate & Advanced SEO | | djo-2836690