What is the point of XML site maps?
-
Given how Google uses Page Rank to pass link juice from one page to the next if Google can only find a page in an XML site map it will have no link juice and appear very low in search results if at all.
The priority in XML sitemaps field also seems pretty much irrelevant to me. Google determines the priority of a page based on the number of inbound links to it. If your site is designed properly the most important pages will have the most links.
The changefreq field could maybe be useful if you have existing pages that are updated regularly. Though it seems to me Google tends to crawl sites often enough that it isn't useful. Plus for most of the web the significant content of an existing page doesn't change regularly, instead new pages are added with new content.
This leaves the lastmod field as being potentially useful. If Google starts each crawl of your site by grabbing the sitemap and then crawls the pages whose lastmod date is newer than its last crawl of the site their crawling could be much more efficient. The site map would not need to contain every single page of the site, just the ones that have changed recently.
From what I've seen most site map generation tools don't do a great job with the fields other than loc. If Google can't trust the priority, changefreq, or lastmod fields they won't put any weight on them.
It seems to me the best way to rank well in Google is by making a good, content-rich site that is easily navigable by real people (and that's just the way Google wants it).
So, what's the point of XML site maps? Does the benefit (if any) outweigh the cost of developing and maintaining them?
-
Thanks Axial,
I'm not convinced it matters much if Google crawls deep pages they wouldn't find through organic links. If the pages aren't linked to they won't have any link juice and therefore won't rank well in SERPs.
The link about using site maps for canonical URLs says or implies you should only put your most important URLs in the sitemap. The sitemap tools I've seen tend to take a kitchen sink approach, which is needed if you are using it to try to get a deeper crawl. Plus there's no way (I see) in a sitemap to specify that page A is the canonical of page B. They simply suggest telling Google about page A (and not page B) in the hopes page A will get more weight than page B. A canonical meta tag on page B pointing to page A is obviously a much better way to deal with canonicals.
Image and video site maps are potentially valuable. I am asking specifically about site maps for pages.
Specifying related content for a given URL, such as different languages, is indeed useful and not something I was aware of. But it is not applicable on most sites and not used on most site maps.
-
Your sitemap.xml will help googlebot crawl deep pages, but it serves other purposes such as:
-
helping Google identify canonical pages: http://support.google.com/webmasters/bin/answer.py?hl=en&answer=139066#3
-
creating sitemaps for video, images, etc.: "you can also use Sitemaps to provide Google with metadata about specific types of content on your site, including video, images, mobile, and News. For example, a video Sitemap entry can specify the running time, category, and family-friendly status of a video; an image Sitemap entry can provide information about an image’s subject matter, type, and license." http://support.google.com/webmasters/bin/answer.py?hl=en&hlrm=fr&answer=156184
-
you can specify alternate content, such as the URL of a translated page: http://support.google.com/webmasters/bin/answer.py?hl=en&answer=2620865
-
and more.
Sometimes working with a sitemap is less risky and maintenance is easier, especially when your CMS is limitative. The 3rd point is a good example. You may also appreciate the centralized approach more from a personnal point of view.
There are good resources on the Google webmaster resources, check them out.
Hope this helps!
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Why is old site not being deindexed post-migration?
We recently migrated to a new domain (16 days ago), and the new domain is being indexed at a normal rate (2-3k pages per day). The issue is the old domain has not seen any drop in indexed pages. I was expecting a drop in # of indexed pages inversely related to the increase of indexed pages on the new site. Any advice?
Algorithm Updates | | ggpaul5620 -
Ecommerce sites 30% drop in organic since the spring
I help manage SEO for a number of large retail websites and we've seen a significant drop in organic traffic (upwards of 30%) since around May 2015. It's likely we were hit my Google's Phantom Quality update, but I don't understand why it had such a big impact. Can anyone explain that Google update in more depth and advise on steps to take to recover from it? Thank you.
Algorithm Updates | | JimLynch0 -
Test site is live on Google but it duplicates existing site...
Hello - my developer has just put a test site up on Google which duplicates my existing site (main url is www.mydomain.com and he's put it up on www.mydomain.com/test/ "...I’ve added /test/ to the disallowed urls in robots.txt" is how he put it. So all the site URLs are content replicated and live on Google with /test/ added so he can block them in robots. In all other ways the test site duplicates all content, etc (until I get around to making some tweaks next week, that is). Is this a bad idea or should I be OK. Last thing I want is a duplicate content or some other Google penalty just because I'm tweaking an existing website! Thanks in advance, Luke
Algorithm Updates | | McTaggart0 -
How can I tell Google two sites are non-competing?
We have two sites, both English language. One is a .ca and the other is a .com, I am worried that they are hurting one another in the search results. I'd like to obviously direct google.ca towards the .ca domain and .com towards the .com domain and let Google know they are connected sites, non-competing.
Algorithm Updates | | absoauto0 -
Implications of removing all google products from site
Is there any data on the implications of removing everything google from a site; analytics, adsense, webmaster tools, sitemaps, etc. Obviously they still have their search data and they say they dont use these other sources of data for ranking information but has anyone actually tried this or is there any existing data on this?
Algorithm Updates | | jessefriedman0 -
How to speed up indexing of my site...
Only 4 out of the 12 pages of my blog/site have been indexed. How can I ensure all the pages get indexed? I'm using a wordpress site, and I also wondered how could I speed the indexing process up (I have submitted a site map) Thanks!
Algorithm Updates | | copywritingbuzz0 -
Why have organic hits to my site suddenly started to fluctuate widely?
Hits to my website have been gradually growing over the last 2 years to a point where there was 1300 organics a day. Recently the hits crashed to 300 a day and stayed there for a couple of days. They then bounced back to 1300 a day for a few more days before crashing down to 300 again. This has happened a few times now. Oscillating between two points. Its like the site is crossing a threshold of something, back and forth. I have looked at different parts of my site and every part is down by approximately the same percentage. Anyone out there have any ideas?
Algorithm Updates | | easymatt0 -
Risks associated with having multiple similar ecom sites together under the same analytics account?
Any downsides to having multiple (similar) eCommerce sites linked to the same Google Analytics account? Traffic splitting or other penalties? I've heard a range of answers from "Yes, traffic was split between my two first-page ranked sites, it was awful" to "no, Google couldn't care less/ they'd be able to tell if your sites were related outside of having them in the same account anyways" Any info would be much apprecaited 🙂 Thanks!
Algorithm Updates | | apo11o1770