Sitemap international websites
-
Hey Mozzers,Here is the case that I would appreciate your reply for: I will build a sitemap for .com domain which has multiple domains for other countries (like Italy, Germany etc.). The question is can I put the hreflang annotations in sitemap1 only and have a sitemap 2 with all URLs for EN/default version of the website .COM. Then put 2 sitemaps in a sitemap index. The issue is that there are pages that go away quickly (like in 1-2 days), they are localised, but I prefer not to give annotations for them, I want to keep clear lang annotations in sitemap 1. In this way, I will replace only sitemap 2 and keep sitemap 1 intact. Would it work? Or I better put everything in one sitemap?The second question is whether you recommend to do the same exercise for all subdomains and other domains? I have read much on the topic, but not sure whether it worth the effort.The third question is if I have www.example.it and it.example.com, should I include both in my sitemap with hreflang annotations (the sitemap on www.example.com) and put there it for subdomain and it-it for the .it domain (to specify lang and lang + country).Thanks a lot for your time and have a great day,Ani
-
Thanks a lot, Tom. I appreciate your time and thorough response. This is exactly what I have been looking for to validate my suggestions and assumptions. I have been careful when implementing each and every annotation in the sitemap, I think I get it right and let's see what will happen.
Cheers,
Annie
-
Hi there,
You can use separate sitemaps along with a sitemap index but when you use you hreflang annotations you must specify all alternates for the URL. or they may not be understood correctly. You’re fine to use a sitemap for all of your content which you don’t wish to add the hreflang tags to, and another for the URLs with hreflang tags.
Just remember to specify every version of each page you mention in your hreflang sitemap along with a <loc>entry all wrapped in a <url>tag:</url></loc>
<url><loc>http://example.com</loc>
<xhtml:link rel="”alternate”" hreflang="”x-default”" href="”http://example.com”">//for users with no version specified
<xhtml:link rel="”alternate”" hreflang="”en”" href="”<a">http://example.com” /> //for English users in any country
<xhtml:link rel="”alternate”" hreflang="”en-us”" href="”<a">http://example.com” /> //us english
<xhtml:link rel="”alternate”" hreflang="”en-gb”" href="”<a">http://example.co.uk” /> //uk english
<xhtml:link rel="”alternate”" hreflang="”it-it”" href="”<a">http://example.it” /> //Italian users in Italy
<xhtml:link rel="”alternate”" hreflang="”it”" href="”<a">http://it.example.com” /> //Italian users anywhere</xhtml:link></xhtml:link></xhtml:link></xhtml:link></xhtml:link></xhtml:link></url>You cannot have one sitemap for hreflang=”en” and another for hreflang=”it” but you can use a separate sitemap on example.it specifying static pages on that domain:
<loc>example.it <loc><loc>example.it/page2</loc></loc></loc>
Your hreflang sitemap on example.it would have the same hreflang tags as the .com, but with the Italian domain specified in <loc>:</loc>
<url><loc>http://example.it</loc>
<xhtml:link rel="”alternate”" hreflang="”x-default”" href="”<a">http://example.com” /> //for users with no version specified
<xhtml:link rel="”alternate”" hreflang="”en”" href="”<a">http://example.com” /> //for English users in any country
<xhtml:link rel="”alternate”" hreflang="”en-us”" href="”<a">http://example.com” /> //us english
<xhtml:link rel="”alternate”" hreflang="”en-gb”" href="”<a">http://example.co.uk” /> //uk english
<xhtml:link rel="”alternate”" hreflang="”it-it”" href="”<a">http://example.it” /> //Italian users in Italy
<xhtml:link rel="”alternate”" hreflang="”it”" href="”<a">http://it.example.com” /> //Italian users anywhere else</xhtml:link></xhtml:link></xhtml:link></xhtml:link></xhtml:link></xhtml:link></url>So, each domain would need its own ‘sitemap 1’ (the hreflang sitemap) and its own sitemap 2 specifying the pages which weren’t in the hreflang sitemap, and its own sitemap index pointing to both sitemaps. Unless you verify both properties under the same WMT account, then you could use a sitemap containing every <loc>from all different sites, along with all their international variations, and reference that 1 international sitemap in your sitemap index for every site – this post will explain multiple domains: https://support.google.com/webmasters/answer/75712</loc>
This webmaster help pages explains about sitemap hreflang implementation: https://support.google.com/webmasters/answer/2620865?hl=en
Hope that helps,
Tom
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Which pages should I index or have in my XML sitemap?
Hi there, my website is ConcertHotels.com - a site which helps users find hotels close to concert venues. I have a hotel listing page for every concert venue on my site - about 12,000 of them I think (and the same for nearby restaurants). e.g. https://www.concerthotels.com/venue-hotels/madison-square-garden-hotels/304484 Each of these pages list the nearby hotels to that concert venue. Users clicking on the individual hotel are brought through to a hotel (product) page e.g. https://www.concerthotels.com/hotel/the-new-yorker-a-wyndham-hotel/136818 I made a decision years ago to noindex all of the /hotel/ pages since they don't have a huge amount of unique content and aren't the pages I'd like my users to land on . The primary pages on my site are the /venue-hotels/ listing pages. I have similar pages for nearby restaurants, so there are approximately 12,000 venue-restaurants pages, again, one listing page for each concert venue. However, while all of these pages are potentially money-earners, in reality, the vast majority of subsequent hotel bookings have come from a fraction of the 12,000 venues. I would say 2000 venues are key money earning pages, a further 6000 have generated income of a low level, and 4000 are yet to generate income. I have a few related questions: Although there is potential for any of these pages to generate revenue, should I be brutal and simply delete a venue if it hasn't generated revenue within a time period, and just accept that, while it "could" be useful, it hasn't proven to be and isn't worth the link equity. Or should I noindex these "poorly performing pages"? Should all 12,000 pages be listed in my XML sitemap? Or simply the ones that are generating revenue, or perhaps just the ones that have generated significant revenue in the past and have proved to be most important to my business? Thanks Mike
Technical SEO | | mjk260 -
Why did my website DA fell down?
Hello, Could you please let me know why might my website's DA have fallen down in merely a week? What might be a reason? I also noticed traffic from google dropped down at the very same week. Will be very thankful for any advise!
Technical SEO | | kirupa0 -
Is sitemap required on my robots.txt?
Hi, I know that linking your sitemap from your robots.txt file is a good practice. Ok, but... may I just send my sitemap to search console and forget about adding ti to my robots.txt? That's my situation: 1 multilang platform which means... ... 2 set of pages. One for each lang, of course But my CMS (magento) only allows me to have 1 robots.txt file So, again: may I have a robots.txt file woth no sitemap AND not suffering any potential SEO loss? Thanks in advance, Juan Vicente Mañanas Abad
Technical SEO | | Webicultors0 -
Website being crawled but not indexed any thoughts?
Hi Everyone,
Technical SEO | | Ant71
I created a new website a few weeks ago www.drivingseaford.co.uk , did a little link citation, links from Google+, submitted to webmaster tools etc but its still not getting indexed. Webmaster tools crawl stats page is showing pages being crawled, no errors. But 0 indexed. http://www.drivingseaford.co.uk/robots.txt is showing User-agent: * Disallow: /wp-admin/ Disallow: /wp-includes/ Im a bit stumped as never had this before!!! Any ideas from you lovely people?? Antony0 -
Internal Links
In OSE, it is reporting that i don't have any internal links to my homepage. In the header on every page is my logo in the top left hand corner which links back to my homepage. Shouldn't this mean then that every page should link to the home page? Similarly, internal pages which link from my main nav aren't showing up as having any internal links in OSE. Any ideas?
Technical SEO | | Santaur0 -
Website Speed
What are the top 10 ways to reduce page loading times and speed up a website?
Technical SEO | | casper4340 -
Xml Sitemap
Hi mozzers, I am about to submit a sitemap for one of my clients via webmaster tools. The issue is that I have way too many urls that I don't want them to be indexed by Google such as testing pages, auto generated pages... Is there way to remove certain URL from the XML sitemap or is this impossible? If impossible, is the only way to control these urls is to "No index" all these pages that i don't want the search engine to see? Thanks Mozzers,
Technical SEO | | Ideas-Money-Art0 -
New website
Hello, How bad is going to be if I change my Joomla website to Wordpress? I can check the 100 best pages and redirect them to the new url with 301 but my website has 424 pages. If is this needs time, how long does it take to be in the same position? Is Google review my new website quickly? What about if I make my services more specific and the main topic is going to be smaller in pages? (Mpre social services pages vs. less pages about the main webdesign topic) I should change my website to WP but I am afraid because now I am in the 2. 🙂 Thanks! Regards,
Technical SEO | | Netkreativ
Misi0