Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Resubmit sitemaps on every change?
-
Hello Mozers,
Our sitemaps were submitted to Google and Bing, and are successfully indexed. Every time pages are added to our store (ecommerce), we re-generate the xml sitemap.
My question is: should we be resubmitting the sitemaps every time their content change, or since they were submitted once can we assume that the crawlers will re-download the sitemaps by themselves (I don't like to assume).
What are best practices here?
Thanks!
-
Great follow up! Thanks for that. :^)
-
For anybody that is interested in knowing, after 10 days of webmaster monitoring, here are our conclusions:
We have 2 sitemaps (2 languages). It took a few days for Google to re-download the mains sitemap, and a few more days to download the secondary sitemap, however, the new sitemaps were both picked up by Google with no intervention on our part.
Bing, however, has still no downloaded either of the updated sitemaps.
Because it's so easy and isn't seen as bad practice, we will be manually re-submitting sitemap updated to both search engines.
Thanks!
-
Ryan gave an excellent answer. Google is using other clues to pick up on new pages I'm not saying don't submit your sitemap I'm just saying add structured data/schema into your store as well. Check your crawl budget and see if your site is eating up too much of it and not being indexed properly by Google.
A simple test to see if something is being blocked is to run your site through https://varvy.com/
If you do not know, I would stress using Deep Crawl or screaming frog SEO spider
Navigation Is often one of many can cause problems where your site will not be crawled correctly.
To determine whether or not you have to crawl budget issue we got each can make independent image sitemaps and index sitemaps as well just to be sure that Google is getting what you want to.
Like Ryan said check out
https://support.google.com/webmasters/answer/183669
Here is a Magento dynamic site map http://i.imgur.com/QKS0bgU.png
validate your sitemap check it for problems
http://tools.seochat.com/tools/site-validator/
https://moz.com/learn/seo/schema-structured-data
http://www.searchmetrics.com/news-and-events/schema-org-in-google-search-results/
https://blog.kissmetrics.com/seo-for-marketplaces-ecommerce/
JSON-LD Microdata
https://builtvisible.com/micro-data-schema-org-guide-generating-rich-snippets/#json
I hope this helps,
Thomas
-
Hello. You can check the Submitted vs Indexed count within Search Console to see whether or not your regenerated sitemap is being picked up already, but resubmitting a sitemap isn't an issue, and fairly easy to do, per Google:
Resubmit your sitemap
- Open the Sitemaps report
- Select the sitemap(s) you want to resubmit from the table
- Click the Resubmit sitemap button.
You can also resubmit a sitemap by sending an HTTP GET request to the following URL, specifying your own sitemap URL: http://google.com/ping?sitemap=http://www.example.com/my_sitemap.xml
Via: https://support.google.com/webmasters/answer/183669 Also from a FAQ in the Webmasters blog they state that, "Google does not penalize you for submitting a Sitemap."
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Automate XML Sitemaps
Quick question, which is the best method that people have for automating sitemaps. We publish around 200 times a day and I would like to make sure as soon as we publish it gets updated in the site map. What is the best method of updating a sitemap so it gets updated immediately after it is published.
Technical SEO | | mattdinbrooklyn0 -
301 Redirects Relating to Your XML Sitemap
Lets say you've got a website and it had quite a few pages that for lack of a better term were like an infomercial, 6-8 pages of slightly different topics all essentially saying the same thing. You could all but call it spam. www.site.com/page-1 www.site.com/page-2 www.site.com/page-3 www.site.com/page-4 www.site.com/page-5 www.site.com/page-6 Now you decided to consolidate all of that information into one well written page, and while the previous pages may have been a bit spammy they did indeed have SOME juice to pass through. Your new page is: www.site.com/not-spammy-page You then 301 redirect the previous 'spammy' pages to the new page. Now the question, do I immediately re-submit an updated xml sitemap to Google, which would NOT contain all of the old URL's, thus making me assume Google would miss the 301 redirect/seo juice. Or do I wait a week or two, allow Google to re-crawl the site and see the existing 301's and once they've taken notice of the changes submit an updated sitemap? Probably a stupid question I understand, but I want to ensure I'm following the best practices given the situation, thanks guys and girls!
Technical SEO | | Emory_Peterson0 -
How would you create and then segment a large sitemap?
I have a site with around 17,000 pages and would like to create a sitemap and then segment it into product categories. Is it best to create a map and then edit it in something like xmlSpy or is there a way to silo sitemap creation from the outset?
Technical SEO | | SystemIDBarcodes0 -
How much will changing IP addresses impact SEO?
So my company is upgrading its Internet bandwidth. However, apparently the vendor has said that part of the upgrade will involve changing our IP address. I've found two links that indicate some care needs to be taken to make sure our SEO isn't harmed: http://followmattcutts.com/2011/07/21/protect-your-seo-when-changing-ip-address-and-server/ http://www.v7n.com/forums/google-forum/275513-changing-ip-affect-seo.html Assuming we don't use an IP address that has been blacklisted by Google for spamming or other black hat tactics, how problematic is it? (Note: The site hasn't really been aggressively optimized yet - I started with the company less than two weeks ago, and just barely got FTP and CMS access yesterday - so honestly I'm not too worried about really messing up the site's optimization, since there isn't a lot to really break.)
Technical SEO | | ufmedia0 -
How much to change to avoid duplicate content?
Working on a site for a dentist. They have a long list of services that they want us to flesh out with text. They provided a bullet list of services, we're trying to get 1 to 2 paragraphs of text for each. Obviously, we're not going to write this off the top of our heads. We're pulling text from other sources and trying to rework. The question is, how much rephrasing do we have to do to avoid a duplicate content penalty? Do we make sure there are changes per paragraph, sentence, or phrase? Thanks! Eric
Technical SEO | | ericmccarty0 -
Which is the best wordpress sitemap plugin
Does anyone have a recommendation for the best xml sitemap plugin for wordpress sites or do you steer clear of plugins and use a sitemap generator then load it up to the root manually?
Technical SEO | | simoncmason0 -
Multiple title and description change is considered spam?
I'm changing title and description several times throught 1-2 months... will Google penalize me? I'm worried that testing several title and description ideas would affect the site rankings on google. Anyone had a problem with that... or can i continue testing as many titles and descriptions as i want?
Technical SEO | | mosaicpro0 -
How to handle sitemap with pages using query strings?
Hi, I'm working to optimize a site that currently has about 5K pages listed in the sitemap. There are not in face this many pages. Part of the problem is that one of the pages is a tool where each sort and filter button produces a query string URL. It seems to me inefficient to have so many items listed that are all really the same page. Not to mention wanting to avoid any duplicate content or low quality issues. How have you found it best to handle this? Should I just noindex each of the links? Canonical links? Should I manually remove the pages from the sitemap? Should I continue as is? Thanks a ton for any input you have!
Technical SEO | | 5225Marketing0