Best practice for site maps?
-
- Is it necessary or good practice to list "static" site routes in the sitemap? I.e. /about, /faq, etc? Some large sites (e.g. Vimeo) only list the 'dynamic' URLs (in their case the actual videos).
- If there are urls NOT listed in a sitemap, will these continue to be indexed?
- What is the good practice for a sitemap index? When submitting a sitemap to e.g. Webmaster tools, can you just submit the index file (which links to secondary sitemaps)? Does it matter which order the individual sitemaps are listed in the index?
-
Crimson offers a great reply and gets a thumbs up from me. I'll just add a bit.
Whether or not you submit a sitemap, Google will visit your site as long as it knows the site exists. If your site offers solid navigation, there is absolutely no need to submit a sitemap. Google will find and crawl all of your pages. If you have coding issues on your site, navigation issues, island pages, etc. then a sitemap is helpful so Google can be aware of these pages it would otherwise not be able to find.
With the above noted, a sitemap is easy to set up and automate. You can pretty much "set it and forget it" so it's still a good practice. About your questions,
1. It's your call. If a page is linked to in your main navigation such as About or FAQ then Google should find it 100% of the time. There is no need to include it in your sitemap but there is no harm either. Either way works.
2. Yes, as per the above as long as Google can find the page it will index them. You can even have horrible coding and navigation and Google may locate your pages if you have earned external links to them from credible sources.
3. Last I checked a sitemap can hold 50k URLs. If your site has more then 50k URLs, then you can break up the sitemaps into smaller files. The advice Crimson shared is correct.
In summary, if you implement all best practices in your site design and do not have any island pages then a sitemap is not needed but it is a nice backup.
-
Hi Shawn,
-
Yes it is good practice to list all urls.
-
As long as these urls are linked to internally then they will be indexed.
-
You only need to submit the index file not every individual sitemap. It doesn't matter which order the individual sitemaps are listed.
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=71453&from=35655&rd=1
Hope this helps.
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
VTEX Infinite Scroll Design: What is On-Page SEO Best Practice?
We are migrating to the VTEX E Commerce platform and it is built on javascript, so there are no <a>tags to link product pages together when there is a long list of products. According to the Google Search Console Help document, "Google can follow links only if they are an</a> <a>tag with an href attribute." - Google Search Console Help document </a>http://support.google.com/webmasters/answer/9112205. So, if there a 1000 products, javascript just executes to deliver more content in order to browse through the entire product list. The problem is there is no actual link for crawlers to follow. Has anyone implemented a solution to this or a similar problem?
Intermediate & Advanced SEO | | ggarciabisco0 -
SEO Best Practices regarding Robots.txt disallow
I cannot find hard and fast direction about the following issue: It looks like the Robots.txt file on my server has been set up to disallow "account" and "search" pages within my site, so I am receiving warnings from the Google Search console that URLs are being blocked by Robots.txt. (Disallow: /Account/ and Disallow: /?search=). Do you recommend unblocking these URLs? I'm getting a warning that over 18,000 Urls are blocked by robots.txt. ("Sitemap contains urls which are blocked by robots.txt"). Seems that I wouldn't want that many urls blocked. ? Thank you!!
Intermediate & Advanced SEO | | jamiegriz0 -
Best SEO practice for multiple languages in website
HI, We would like to include multiple languages for our global website. What's the best practice to gain from UI and SEO too. Can we have auto language choosing website as per browsing location? Or dedicated pages for important languages like www.website.com/de for German. If we go for latter, how about when users browsing beside language page as they will be usually in English
Intermediate & Advanced SEO | | vtmoz0 -
Can I add Title/Description tags to site map
I have started working on a website that it written in JAVA. It has 26 URL's But because of the way it is written it is all shown on the home page code and does not have the ability to add unique title and description tags. Is there a work around for SEO on websites like this aside from adding content? I was wondering if there is a way to submit a sitemat with title and description tags. Any advice? Chris.K
Intermediate & Advanced SEO | | CKerr0 -
Slug best practices?
Hello, my team is trying to understand how to best construct slugs. We understand they need to be concise and easily understandable, but there seem to be vast differences between the three examples below. Are there reasons why one might be better than the others? http://www.washingtonpost.com/news/morning-mix/wp/2014/06/20/bad-boys-yum-yum-violent-criminal-or-not-this-mans-mugshot-is-heating-up-the-web/ http://hollywoodlife.com/2014/06/20/jeremy-meeks-sexy-mug-shot-felon-viral/ http://www.tmz.com/2014/06/19/mugshot-eyes-felon-sexy/
Intermediate & Advanced SEO | | TheaterMania0 -
Site: inurl: Search
I have a site that allows for multiple filter options and some of these URL's have these have been indexed. I am in the process of adding the noindex, nofollow meta tag to these pages but I want to have an idea of how many of these URL's have been indexed so I can monitor when these have been re crawled and dropped. The structure for these URL's is: http://www.example.co.uk/category/women/shopby/brand1--brand2.html The unique identifier for the multiple filtered URL's is --, however I've tried using site:example.co.uk inurl:-- but this doesn't seem to work. I have also tried using regex but still no success. I was wondering if there is a way around this so I can get a rough idea of how many of these URL's have been indexed? Thanks
Intermediate & Advanced SEO | | GrappleAgency0 -
Potential problems with my site
Dear Mozzers I hope you can help me with the following problems: My site is up and running for a year now and may be there has been problem with the homepage, because it ranks on first page for a competitive keyword on Google.com and Google.com.au only, however in other countries it just shows up as internal page and does not rank well. google.com/Google.com.au: homepage ranks top 10 (example.com)
Intermediate & Advanced SEO | | SteveTran2013
Other countries (.co.uk, .ca..ect) an internal page shows up, example.com/internalpage.html - shows on page 3-4. I can not find the homepage of example.com anywhere around top 1000. Can you please tell me what are the potential problems. Thank you very much. BR/Tran0 -
Site speed tests
In webmaster tools my site is showing that it is taking longer and longer to load, and it has now doubled. Is there a way to check which pages are the problem? The site is quite large so I can't check them one at a time.
Intermediate & Advanced SEO | | EcommerceSite0