Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Best practice for URL - Language/country
-
Hi,
We are planning on having our website localized into more languages. We already have an English and German version. The German version is currently a sub-domain:
www.example.com --> English version
de.example.com --> German version
Is this recommended? Or is it always better to have URLs with language prefixes such a:
Which is a better practice in terms of SEO?
-
Hi Peter,
Both really good answers to your questions above but maybe it would be good to give you some further pointing in the right direction. Perhaps you could answer the questions below and I can give you my personal opinion on which method would be best:
-
will you be putting an equal amount of marketing (content, PR, etc.) into the Spanish version for example compared with English?
-
are you able to offer fully localised service eg, Spanish customer service, Spanish sales team etc.?
-
is your company well-known globally?
It's important not to also forget that another option is using ccTLDs (eg, .co.uk, .com.au). These give the highest signal to search engines about the country being targeted and also importantly make you look more "local" which can do wonders for increasing conversion rate in countries where your company is not well-known.
-
-
I think that Tom gave you one of the best answers possible.
However I hope this helps your site structure should be very similar to one contained in the two URL's
If I may add a little bit of information that I thought was helpful
- https://support.google.com/webmasters/answer/189077?hl=en
- https://www.deepcrawl.com/knowledge/best-practice/hreflang-101-how-to-avoid-international-duplication/
WHERE TO ADD YOUR HREFLANG TAGS
You can add hreflang tags to your sitemaps, in the HTTP response headers, or on the page itself.
IN YOUR SITEMAPS
The best place to add hreflang is in your sitemap as including them in the headers or on the page adds weight to every single page request.
The following example will inform Google about the English version from the German version of the website:
<url> <loc>http://www.example.com/deutsch/</loc></url>
<xhtml:link< span=""> rel=”alternate” hreflang=”en” href=”http://www.example.com/english/” /> <xhtml:link < span="">rel=”alternate” hreflang=”de” href=”http://www.example.com/deutsch/” /></xhtml:link <></xhtml:link<>
This method would need to be repeated in full for every page on the site and for all the international websites.
IN YOUR HEADERS AND HTML
Hreflang tags can also be added to the HTTP header:
Link: http://www.example.com/english/; rel=”alternate”; hreflang=”en” Link: http://www.example.com/deutsch/; rel=”alternate”; hreflang=”de”
Or in the tag in the HTML:
http://www.example.com/english/” /> http://www.example.com/deutsch/
& because you will be creating a new site
https://www.candidsky.com/blog/the-seo-2015-guide-to-website-migration/
it would come down to your backlink profile if it were me I would use
Moz open site Explorer, Majestic, Ahrefs and Google Webmaster tools to determine whether or not I will be receiving a enough Backlinks for a subdomain or separate TLD otherwise I would use a subfolder and an extremely fast method of hosting the site Fastly is excellent or many other great methods as well.
Hope this helps,
Tom
PS use
http://hreflang.ninja/ to check
-
Hi Peter
Both are viable options.
I'd highly recommend going through Aleyda Solis' international SEO posts here on the Moz blog. They can teach how to prepare for international SEO, how to approach site structure and how to generate relevant code and hreflang tags.
Here is her international SEO checklist
Here is her Hreflang blog post and generator tool
And 40 tools to help advance your international SEO
They're great reading and nothing that I'd be able to do add to, so I hope this helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Sudden Indexation of "Index of /wp-content/uploads/"
Hi all, I have suddenly noticed a massive jump in indexed pages. After performing a "site:" search, it was revealed that the sudden jump was due to the indexation of many pages beginning with the serp title "Index of /wp-content/uploads/" for many uploaded pieces of content & plugins. This has appeared approximately one month after switching to https. I have also noticed a decline in Bing rankings. Does anyone know what is causing/how to fix this? To be clear, these pages are **not **normal /wp-content/uploads/ but rather "index of" pages, being included in Google. Thank you.
Technical SEO | | Tom3_150 -
Exact Match Domain & Title Tag / URL
I currently own an exact match domain for my keyword. I have it set up with multiple pages and also a blog. The home page essentially serves as a hub and contains links to all the pages and the blog. My targeted keyword is on its own page and I made the title tag the same as my keyword. As an example the URL for my targeted post looks like this: benefitsofrunningshoes.com/benefits-of-running-shoes I have solid, non-spammy content and clean whitehat earned backlinks directing to that specific page. My concern right now is that the URL looks kinda spammy. The website has been live for about a week and the home page ranks well enough but my targeted page is no where to be found. (it does show up if I manually search via search command "site:benefitsofrunningshoes.com"). I'm wondering if it is acceptable to use the exact keyword in title tag / page url if it is also in the domain as an EMD? Should I change the title tag and leave the URL in? Or should I completely change the title tag and URL and 301 redirect to the new page? I appreciate any help!
Technical SEO | | Kusanagi170 -
Numbers in URL
Hey guys! Need your many awesome brains. 🙂 This may be a very basic question but am hoping you can help me out with some insights beyond "because Google says it's better". 🙂 I only recently started working with SEO, and I work for a SaaS website builder company that has millions of open/active user sites, and all our user sites URLs, instead of www.mydomainname.com/gallery or myusername.simplesite.com/about, we use numbers, so www.mysite.com/453112 or myusername.simplesite.com/426521 The Sales manager has asked me to figure out if it will pay off for us in terms of traffic (other benefits?) to change it from the number system to the "proper" and right way of setting up these URLs. He's looking for rather concrete answers, as he usually sits with paid search and is therefore used to the mindset of "if we do x it will yield us y in z months". I'm finding it quite difficult to find case studies/other concrete examples beyond the generic, vague implication that it will simply be "better" (when for example looking at SEO checklists and search engine guidelines). Will it make a difference? How so? I have to convince our developers of the importance and priority of this adjustment, or it will just drown in the many projects they already have. So truly, any insights would be so very welcome. Thank you!
Technical SEO | | michelledemaree2 -
Disallow: /404/ - Best Practice?
Hello Moz Community, My developer has added this to my robots.txt file: Disallow: /404/ Is this considered good practice in the world of SEO? Would you do it with your clients? I feel he has great development knowledge but isn't too well versed in SEO. Thank you in advanced, Nico.
Technical SEO | | niconico1011 -
WordPress - How to stop both http:// and https:// pages being indexed?
Just published a static page 2 days ago on WordPress site but noticed that Google has indexed both http:// and https:// url's. Usually I only get http:// indexed though. Could anyone please explain why this may have happened and how I can fix? Thanks!
Technical SEO | | Clicksjim1 -
Robots.txt to disallow /index.php/ path
Hi SEOmoz, I have a problem with my Joomla site (yeah - me too!). I get a large amount of /index.php/ urls despite using a program to handle these issues. The URLs cause indexation errors with google (404). Now, I fixed this issue once before, but the problem persist. So I thought, instead of wasting more time, couldnt I just disallow all paths containing /index.php/ ?. I don't use that extension, but would it cause me any problems from an SEO perspective? How do I disallow all index.php's? Is it a simple: Disallow: /index.php/
Technical SEO | | Mikkehl0 -
Old URL redirect to New URL
Alright I did something dumb a year a go and I'm still paying for it. I changed my hyphenated URL to the non-hyphenated version when I redesigned my website. I say it was dumb because I lost most of my link juice even though I did 301 redirects (via the htaccess file) for almost all of the pages I could find in Google's index. Here's my problem. My new site took a huge hit in traffic (down 60%) when I made the change and even though I've done thousands of redirects my old site is still showing up in the SERPS and send much if not most of my traffic. I don't want to take the old site down in fear it will kill all of my traffic. What should I do? Is there a better method I should explore then 301 redirects? Could the other site be affecting my current rank since it's still there? (FYI...both sites are built on the WP platform). Any help or ideas are greatly appreciated. Thank you! Joe
Technical SEO | | kaje0 -
What is the best method to block a sub-domain, e.g. staging.domain.com/ from getting indexed?
Now that Google considers subdomains as part of the TLD I'm a little leery of testing robots.txt with something like: staging.domain.com
Technical SEO | | fthead9
User-agent: *
Disallow: / in fear it might get the www.domain.com blocked as well. Has anyone had any success using robots.txt to block sub-domains? I know I could add a meta robots tag to the staging.domain.com pages but that would require a lot more work.0