Using Sitemap Generator - Good/Bad?
-
Hi all
I recently purchased the full licence of XML Sitemap Generator (http://www.xml-sitemaps.com/standalone-google-sitemap-generator.html) but have yet used it.
The idea behind this is that I can deploy the package on each large e-commerce website I build and the sitemap will be generated as often as I set it be and the search engines will also be pinged automatically to inform them of the update. No more manual XML sitemap creation for me!
Now it sounds great but I do not know enough about pinging search engines with XML sitemap updates on a regular basis and if this is a good or bad thing?
Can it have any detrimental effect when the sitemap is changing (potentially) every day with new URLs for products being added to the site?
Any thoughts or optinions would be greatly appreciated.
Kris
-
It would certainly not have any impact on the existing rankings and crawl rate. infact the crawl rate would improve
-
Hi Khem
Yes I fully understood your response, thank you.
We always do sumbit a sitemap file for every site we build however we never really update the sitemap from there on, we tend to leave the search engines to crawl the site in order to find new pages or detect pages that have been removed. I assume this is a normal practice for many other developers out there?
We always do a 301 redirect for pages which have been un-published such as old products and categories.
My main concern was actually creating a new sitemap file every day and if this would have any effect on the existing rankings, or crawl rate of the site. I guess not!
Kris
-
Well to answer you question. Yes, you should always use xml sitemap also submit it with search engines. It helps search engines to access all the pages of your websites. If fact you can even tell search engines about your most important and less important pages.
It also enables you to tell Search Engines about the content update frequency so that search engine could crawl those again which you update daily/ weekly.
Furthermore, there is no problem if you update the XML file daily as long as you're not removing pages. However, if you need to remove pages, keep them in sitemap for at least one week and redirect old pages to new ones.
Hope I was able to understand your question and answered properly
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
I have duplicate content but // are causing them
I have 3 pages duplicated just by a / Example: https://intercallsystems.com/intercall-nurse-call-systems**//**
Technical SEO | | Renalynd
https://intercallsystems.com/intercall-nurse-call-systems**/** What would cause this?? And how would I fix it? Thanks! Rena0 -
Date in permalinks. Bad?
Hello! I have a recipe website with over 1000 posts. Currently I have the month and year in the permalink that everyone is hinting off to me is bad. On the same front people tell me if I change the permalinks to just the post name it's going to significantly slow down my site. I'm torn on this one about changing. From Google's standpoint is it better to change to the post name and if so should I be fearing I'm going to run into trouble with the change? Any suggestions you have would be appreciated. Thanks!!!
Technical SEO | | Rich-DC1 -
Sitemap
Hi, I am setting up a new sitemap for our website. the website contains about 8000 - 10.000 pages. Of wich are 6000 productpages. I have 10 categories, about 80 sub-catagories and about 400 sub-sub categories ( these ar my most important landingpages) At this moment our sitemap is only 1 MB. From that point of view 1 sitemap will be enough. But can i take SEO advantage by splitting this sitemap in 10 categories? Or are there other ways to set it up for a better SEO? Thanks!
Technical SEO | | Leonie-Kramer0 -
Pagination/Crawl Errors
Hi, Ive only just joined SEO moz and after they crawled my site they came up with 3600 crawl errors mostly being duplicate content and duplicate urls. After researching this it soon became clear it was due to on page pagination and after speaking with Abe from SEO mozhe advised me to take action by getting our developers to implement rel=”next” & rel=”prev” to review. soon after our developers implemented this code ( I have no understanding of this what so ever) 90% of my keywords I had been ranking for in the top 10 have dropped out the top 50! Can anyone explain this or help me with this? Thanks Andy
Technical SEO | | beck3980 -
Auto generated pages
Hi, I have two sites showing (crawl report from SEOMoz.org) extremely high numbers of duplicate titles and descriptions (e.g., 33,000). These sites have CMSs behind them and so the duplicate titles, etc., are a result of auto-generated pages. What is the best way to address these problems? Thanks! David
Technical SEO | | DWill0 -
Robots.txt to disallow /index.php/ path
Hi SEOmoz, I have a problem with my Joomla site (yeah - me too!). I get a large amount of /index.php/ urls despite using a program to handle these issues. The URLs cause indexation errors with google (404). Now, I fixed this issue once before, but the problem persist. So I thought, instead of wasting more time, couldnt I just disallow all paths containing /index.php/ ?. I don't use that extension, but would it cause me any problems from an SEO perspective? How do I disallow all index.php's? Is it a simple: Disallow: /index.php/
Technical SEO | | Mikkehl0 -
Using a non-visible H1
I have a developer that wants to use style="text-indent:-9999px" to make the H1 non-visible to the user. Being the conservative person I am, I've never tried this before and worry that Search Engines may think this is a form of cloaking. Am I worrying about nothing? And apologies if it's already been covered here. I couldn't find it. Thanks in advance!!!!
Technical SEO | | elytical0 -
I have a site that has both http:// and https:// versions indexed, e.g. https://www.homepage.com/ and http://www.homepage.com/. How do I de-index the https// versions without losing the link juice that is going to the https://homepage.com/ pages?
I can't 301 https// to http:// since there are some form pages that need to be https:// The site has 20,000 + pages so individually 301ing each page would be a nightmare. Any suggestions would be greatly appreciated.
Technical SEO | | fthead90