Replacing a site map
-
We are in the process of changing our folder/url structure. Currently we have about 5 sitemaps submitted to Google.
How is it best to deal with these site maps in terms of either (a) replacing the old URLs with the new ones in the site map and (b) what affect should we have if we removed the site map submission from the Google Webmaster Tools console.
Basically we have in the region of 20,000 urls to redirect to the new format, and to update in the site map.
-
Another thought might be to place a noindex on the new pages to start with and as we migrate and 301 redirect the old to the new remove the noindex on the new pages ?
That can work but it's not an approach i would use. It seems like a lot of extra work, you run the risk of forgetting to remove the noindex tag on some pages, and also you may wind up not having pages properly indexed for a month.
If you publish a page today, Google may crawl the new page and see the noindex tag. You can then remove the noindex tag but Google may not recrawl the page for some time leaving your site without an indexed page.
As part of the process of publishing the page, I would 301 the old URL to the new URL immediately.
-
Another thought might be to place a noindex on the new pages to start with and as we migrate and 301 redirect the old to the new remove the noindex on the new pages ?
Thoughts ??
-
since the site has over 10,000 pages we need to make sure all redirects etc are set-up before we go live with the new URLs ?
Whether your site has 10 pages or a million pages you should ensure all internal links work without the need for redirection. Any old external links should be redirected to the correct page on your site if one exists. Otherwise you can allow the URL to 404 if there is not a current equivalent page.
Set up your site's 404 page so users are offered a basic "page not found" message along with your site's navigation and a search function. You should set up a log to track which URLs are generating 404 errors.
Prior to launching the site run a crawl diagnostic to help ensure nothing has been missed.
-
Perfect thanks. Just one final question, since the site has over 10,000 pages we need to make sure all redirects etc are set-up before we go live with the new URLs ?
What is the best way to go forward with regards launching the site ?
Should we launch the new pages and then go through the URLs redirecting them ?
Thoughts please ??
-
We are changing our site structure for two main reasons
-
Ease of functionality, and having the ability to target friendly URLs suitable for SEO
-
Plus we've a new CMS, that allows this custom written URLs
The current structure has too many folders that are too deep, and is becoming too un-manageable. The new CMS gives us totally control from one control panel.
I understand that we will loose some PR, but believe it will be for the better of the site and user experience.
-
-
We are changing our site structure for two main reasons
-
Ease of functionality, and having the ability to target friendly URLs suitable for SEO
-
Plus we've a new CMS, that allows this custom written URLs
The current structure has too many folders that are too deep, and is becoming too un-manageable. The new CMS gives us totally control from one control panel.
I understand that we will loose some PR, but below it will be for the better of the site and user experience.
-
-
Big question: are you changing folder/url structure for aesthetics or functionality? Often times it's not worth making such a large change in hopes of getting some SEO-friendly URL's, as the weight on SEO-friendly URL's isn't what it once was. And the headache involved, as well as the inevitable loss in traffic, is quite often not worth it at all.
With that said, refresh your entire sitemap with the new URL's once they are made. Remove all old urls.
IMPORTANT: setup 301 redirects, either using .htaccess or PHP (or whatever language your site uses), to redirect all old urls to the respective new urls. You will lose a fair chunk of PR during this change, but if you feel your site will benefit greatly from a structure change, then you will be willing to take the hit.
Don't leave any redirect un-turned. Then, you'll just have to wait it out while Google re-indexes your entire site trying to figure out your new url structure. Could take a week, could take months. All depends on what Google has valued your site as. For example, if CNN changed their entire URL structure, they probably would miss a beat. Smaller websites tend to take much larger hits in the SERP's.
So, just be sure it's a necessary action, trust me. And don't ever remove those 301's from your .htaccess as you never know what Google still has in their index for your site.
-
A sitemap should be a link representation of your site. It should contain a link to every page you wish to be included in Google's index.
How is it best to deal with these site maps in terms of either (a) replacing the old URLs with the new ones in the site map
Just make the switch. If a page no longer exists on your site, remove the link. If you create a new page on your site, add the link.
what affect should we have if we removed the site map submission from the Google Webmaster Tools console
For the most part, none. During the next crawl Google would look for your sitemap at your root address: www.mydomain.com/sitemap.xml. Google will also check your robots.txt file for a path to your sitemap file. If a sitemap is not located, it will crawl your site normally.
The primary purpose of a sitemap is to allow Google to become aware about new pages on your site it otherwise might not find. If your site offers solid navigation, a site map is not necessary at all.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
PortfolioID urls appearing in my wordpress site- what to do?
Hey guys, Hoping someone may have some advice on a wordpress site. Most of their URL's are duplicates due to a PortfolioID appearing in the URLs causing a duplicate title tags
Technical SEO | | Swanny_s
It's the same page but it's being flagged as duplicate. Would you remove the portfolioID url or 301 redirect? Many thanks
Simon0 -
Duplicate Content within Site
I'm very new here... been reading a lot about Panda and duplicate content. I have a main website and a mobile site (same domain - m.domain.com). I've copied the same text over to those other web pages. Is that okay? Or is that considered duplicate content?
Technical SEO | | CalicoKitty20000 -
What should the combined domain rank of 2 sites?
We have just combined 2 sites by redirecting a site with a domain ranking of 35 to a site with a domain ranking of 27 (this is because the we want to use the address with a ranking of 27) After a week I now have 1 domain with a ranking of 26?? So my questions are; Should the ranking have increased already? What should I expect the ranking to increase to if at all? Is there something I could have done wrong when transferring? Thanks
Technical SEO | | benj450 -
Google having trouble accessing my site
Hi google is having problem accessing my site. each day it is bringing up access denied errors and when i have checked what this means i have the following Access denied errors In general, Google discovers content by following links from one page to another. To crawl a page, Googlebot must be able to access it. If you’re seeing unexpected Access Denied errors, it may be for the following reasons: Googlebot couldn’t access a URL on your site because your site requires users to log in to view all or some of your content. (Tip: You can get around this by removing this requirement for user-agent Googlebot.) Your robots.txt file is blocking Google from accessing your whole site or individual URLs or directories. Test that your robots.txt is working as expected. The Test robots.txt tool lets you see exactly how Googlebot will interpret the contents of your robots.txt file. The Google user-agent is Googlebot. (How to verify that a user-agent really is Googlebot.) The Fetch as Google tool helps you understand exactly how your site appears to Googlebot. This can be very useful when troubleshooting problems with your site's content or discoverability in search results. Your server requires users to authenticate using a proxy, or your hosting provider may be blocking Google from accessing your site. Now i have contacted my hosting company who said there is not a problem but said to read the following page http://www.tmdhosting.com/kb/technical-questions/other/robots-txt-file-to-improve-the-way-search-bots-crawl/ i have read it and as far as i can see i have my file set up right which is listed below. they said if i still have problems then i need to contact google. can anyone please give me advice on what to do. the errors are responce code 403 User-agent: *
Technical SEO | | ClaireH-184886
Disallow: /administrator/
Disallow: /cache/
Disallow: /components/
Disallow: /includes/
Disallow: /installation/
Disallow: /language/
Disallow: /libraries/
Disallow: /media/
Disallow: /modules/
Disallow: /plugins/
Disallow: /templates/
Disallow: /tmp/
Disallow: /xmlrpc/0 -
Webmaster Tools Links To Your Site
I logged onto webmaster tools today for my site and the section 'Links to Your Site' is showing no data. Also if I search using link:babskibaby.com it only shows 1 link. My site had been showing 500+ links previously. Does anyone know why this is?
Technical SEO | | babski0 -
One large site or a few microsites?
Hi, I have a client who runs a professional expo company and wants to redo his website. Right now he has one website that has the following sections: expo company info, wedding show info, electrical wire show info, fishing show info. My question is, when I rebuild the site would it be better to do one site or four microsites?
Technical SEO | | JohnWeb120 -
What to include on a sitemap for a huge site?
I have a very large site and I'm not sure what all to include on the sitemap page. We have categories such as items1, items2 and in the items1 category are 100 vendors with their individual vendor pages. Should I link all 100 vendor pages on the sitemap or just the main items1 category?
Technical SEO | | CFSSEO0 -
E-Commerce site and blogs
We have e-Commerce site and an official blog to give advice about our products. This blog exists under our domain. Usually we build links directly to our site. Recently our ranking started going down. Also, we have been experiencing backlash for spam based on our link building (we are working on this, including a change of staff,but we cannot be sure that this will not happen again). This backlash has come through our social networking outlets (Facebook) in the form of very negative posts to our pages. One of our "SEOs" has devised a plan to use secondary blogs which we would start building links for. This blog would contain links back to our website. The idea is that the blog acts as a gate in a sense, in this way backlash is either posted on the blog or is directed at the blog. Also, we would be attempting to raise the page authority of these secondary blogs so in essence they act as high page authority links back to our website. The concern is that these secondary blogs may undermine the legitimacy of the official primary blog, which is still in its early stages as far as ranking and authority goes. Also, we are concerned that this technique would further undermine the legitimacy of the website itself by creating a larger "spam-like" presence, since visitors may see through the use of the secondary link through blogs.
Technical SEO | | ctam0