Hreflang for multple countries but single language
-
I'm working on a site that has implemented hreflang.
The site is all in English but has slight differences and breaks down to separate domains, so
the hreflang implementation targets specific countries per domain.
This was tested using http://flang.dejanseo.com.au but Webmaster Tools has thrown up errors. For example,
**URLs for your site and alternative URLs in 'en_GB' that do not have return tags. **But every page refers back.Any ideas what's going wrong?
-
The only valid redirection is the one based on user browser, not IP, IMHO.
However, if you want to use that kind of redirection for the home page (not the others pages), then that redirection should be working the first time, so that users can eventually choose to go to another version of the site they prefer (i.e.: I live in Spain and I go to domain.eu. When I travel in the US, I still want to go to domain.eu, not always being pushed to use domain.com).
Moreover, doing that, you will let crawlers to discover also the others version even though they were redirected to the one corresponding to their IP the first time. In other words, Googlebot (Mountain View, USA, IP), the first time will go always to domain.com, but once there it will able to discover also domain.eu and domain.co.uk from the versions selector, and it won't be redirected again to domain.com.
Said all this, the fact is that you want to target a political area (EU) with domain.eu, the world with domain.com and the UK with domain.co.uk.
This desire to target three different kind of geographies complicates everything :-).
The only solution I see is:
-
Domain.com set up as "global" >> hreflang="en". All users using English will see it in the SERPs despite of their location but in these cases (see point 2 and 3);
-
Domain.eu set up as the site for the European countries >> hreflang="en-ES" - hreflang="en-IT" - hreflang="en-DE" and so on.
-
Domain.co.uk set up as the site for Great Britain >> hreflang="en-GB".
Doing this and implementing the hreflang in the canonical URLs of the sites and referencing only canonical URLs of the others sites (apart having the self-referral hreflang), then you should be safe.
However, remember that with those hreflang, people searching in another language than English will never (or almost never) see your sites but for brand name searches or very specific brand + product queries.
Therefore, I am still of the idea that having only English websites for targeting the world means missing a huge business opportunity.
-
-
Hi Michael,
The tool crawls with the user agent "Hreflang.org Testing Tool - Desktop version". It does not impersonate Googlebot. It does not follow 301 redirects because if you use Hreflang to point to another page, you should link to the canonical version of that page. A 301 redirect indicates an error.
My opinion is that forced, IP-based redirection is not a good idea because all URLs must be accessible from all locations. Google recently (about a year ago?) started crawling from other countries but you still want all URLs to be accessible from everywhere. The point of Hreflang is to indicate links across URLs; so hreflang makes IP-based redirection unnecessary.
If you really want to do IP-based redirection, do it on the x-default version of the page. And use the x-default version solely for redirection. For example, don't have your English version as the x-default and then redirect non-English users to other versions. If you do that, Germany-based crawlers won't ever be able to access your English URL.
-
Hi Nikhilesh,
as a matter of interest is your tool crawling as a google bot, rather than some IP? We have now set to deliver TLD depending on IP but the tool reports the following for the hreflangs not on the same domain
- Could not load this page. Server responded with HTTP Status code 301. Expected HTTP response code is 200 OK.
-
thanks. I'm taking a look.
-
I believe what's happening with your site is the same thing as what happened here on Google webmaster forums. JohnMu replied in that thread to say that because you have the same content on different sites (it might say en-US or en-GB but there is no detectable language customization for country). Google considers this duplicate content and starts ignoring some of your pages. And when some pages get ignored, the hreflang return tags from those pages go missing.
I blogged about this phenomenon in detail here.
Of course, it's also possible that your Hreflang markup has errors that the dejan seo tool did not detect. You can try the tool at hreflang.org to get a second opinion.
-
This is a bit difficult. All 3 domains are bringing in some good diverse country traffic as is, but Google has said it will now ignore hreflang until resolved.
So for example, .com brings in traffic and sales from Chile, Mexico and Bahamas. So if we removed the 'global' for .com are we not risking dropping the traffic for these locations. Or will Google ignore anyway and serve? There are too many countries getting traffic to implement each one as hreflang and remove the global en, it would be massive.
But having said that if the server was setting the TLD to be served depending on location it would then be adequate to just use the 'global' hreflang's to hreflang="en"? Is that right?
-
You should consider using either one subfolder hreflang or strongly considering wire going into these markets for instance Germany would be best for .de if you are just targeting a European country that speaks English for only English-speaking countries throughout the world you may be best off with A geotld
i agree with Gianluca.
-
If your .eu domain targets English in all European countries but UK (and you are targeting the USA with the .com), then you should not use the hreflang:
because that annotation says to Google to show the .eu domain to all the people using English independently from where they are, but in the UK ("show domain.co.uk") and USA ("show domain.com"). In other words, you .eu site would be visible to Canadian, Australians or even people in Greenland using English.
If that is also the meaning of the .eu domain, than that's fine, but if it is meant just for European countries, then better not having it.
Moreover, if the .eu is meant as "global", then it is useless to add all the hreflang like "en-it", "en-es" and so on, because the hreflang="en" would be enough.
However, I hope you are considering the .eu in English for targeting countries like Italy, Spain, France, Germany et al as temporary, because - let me tell you - you are not going to see lot of traffic from European not English speaking countries. Why? Because we don't search in English, neither we have our browser set up to use English as main language
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Country and Language tags-Running an SEO audit on a site that definitely has more than one language, but nothing is pulling up. I don't quite understand href lang or how to go about it. HELP HELP!
Ran an SEO audit and I don't really understand country and language tags. For example, sony.com definitely has more than one language, but how do I seo check href lang ? Do I inspect the page? etc?
Technical SEO | | Mindgruver0 -
Will putting a one page site up for all other countries stop Googlebot from crawling my UK website?
I have a client that only wants UK users to be able to purchase from the UK site. Currently, there are customers from the US and other countries purchasing from the UK site. They want to have a single webpage that is displayed to users trying to access the UK site that are outside the UK. This is fine but what impact would this have on Google bots trying to crawl the UK website? I have scoured the web for an answer but can't find one. Any help will be greatly appreciated. Thanks 🙂
Technical SEO | | lbagley0 -
Google tries to index non existing language URLs. Why?
Hi, I am working for a SAAS client. He uses two different language versions by using two different subdomains.
Technical SEO | | TheHecksler
de.domain.com/company for german and en.domain.com for english. Many thousands URLs has been indexed correctly. But Google Search Console tries to index URLs which were never existing before and are still not existing. de.domain.com**/en/company
en.domain.com/de/**company ... and an thousand more using the /en/ or /de/ in between. We never use this variant and calling these URLs will throw up a 404 Page correctly (but with wrong respond code - we`re fixing that 😉 ). But Google tries to index these kind of URLs again and again. And, I couldnt find any source of these URLs. No Website is using this as an out going link, etc.
We do see in our logfiles, that a Screaming Frog Installation and moz.com w opensiteexplorer were trying to access this earlier. My Question: How does Google comes up with that? From where did they get these URLs, that (to our knowledge) never existed? Any ideas? Thanks 🙂0 -
Meta tags in Single Page Apps
Since the deprecation of the AJAX Crawling Scheme back last October I am curious as to when Googlebot actually reads meta tag information from a page. We have a website at whichledlight.com that is implemented using emberjs. Part of the site is our results pages (i.e. gu10-led-bulbs). This page updates the meta and link tags in the head of the document for things like canonicalisation and robots, but can only do so after the page finishes loading and the JavaScript has been run.When the AJAX crawling scheme was still in place we were able to prerender these pages (including the modified meta and link tags) and serve these to Googlebot. Now Googlebot no longer uses these prerendered snapshots and instead is sophisticated enough load and run our site.So the question I have is does Googlebot read the meta and links tags downloaded from the original response or does it wait until the page finishes rendering before reading them (including any modifications that have been performed on them)
Technical SEO | | TrueluxGroup1 -
Roger returned single page
A search query in Google shows 156 indexed results. Yet Roger has completed a crawl and returned a single page. The website is table based, but there appear to be no redirects or Javascript blocking bots so I'm unsure why Roger has under delivered. The problem is not unique to Roger. I ran the site through Screaming Frog's desktop crawler and it also returned a single page. I'm wondering if there is something in the site code I don't know to look for that is stopping Roger & Screaming Frog from crawling the site. Appreciate any insights you can offer. PS. I've read https://seomoz.zendesk.com/entries/409821-Why-Isn-t-My-Site-Being-Crawled-You-re-Not-Crawling-All-My-Pages- and don't think these suggested causes apply.
Technical SEO | | NicDale0 -
Multi language blog
Does google consider it to be spam by having multiple languages posts on a blog such as English, Arabic, spanish blog posts.
Technical SEO | | MozAddict0 -
Country domain: Seo for other languages
Hi, I have an italian domain (.it) for an italian hotel, it is an old authoritative domain (1997) and it is well optimized for the keywords that include the city the hotel is in, now the page is decently positioned in Google Italy. There are many problems to have the same rank for German version (in google.de, google.at). The German version is in the /de folder. The hotel has another .com domain, much less authoritative (2007), in a German server, but it was and is only a simple redirect 301 (by code) to the German version in the .it domain. (obviously the rank for this domain is almost nonexistent). Do you have any suggestion? Thank you.
Technical SEO | | depi0 -
301 Single Page Redirects in IIS7?
Hey all -- I am working with a client, getting ready to make a full domain level change to a brand new domain. The existing domain has solid domain importance and trust, and the home page has a 5/10 GPR, so the transfer of all existing link juice is very important. Of course, I will be utilizing 301's to permanently redirect all existing pages to their new permanent homes. It will be a 1-1 structure, which I know is also best when possible. My question comes in specific to IIS. There is a wealth of information out there on the net regarding implementing permanent 301's using Apache and .htaccess, but nada when it comes to doing it in IIS7, which is what the client is using. For instance, today I am seeking to help them redirect 2 single pages to new destinations within the same domain, just diffferent folders. When you open up the IIS7 Control Panel (yes, with full Admin access), you can navigate to the directory, but the individual pages that I am looking to redirect with 301's do not show in IIS7, so you can't just right click on each page and choose "A redirection to a URL," etc. Any help on exactly how to redirect a single page using a permanent 301 in IIS 7 would be huge! Thanks guys!
Technical SEO | | Bandicoot0