Mass URL changes and redirecting those old URLS to the new. What is SEO Risk and best practices?
-
Hello good people of the MOZ community,
I am looking to do a mass edit of URLS on content pages within our sites. The way these were initially setup was to be unique by having the date in the URL which was a few years ago and can make evergreen content now seem dated. The new URLS would follow a better folder path style naming convention and would be way better URLS overall.
Some examples of the **old **URLS would be
https://www.inlineskates.com/Buying-Guide-for-Inline-Skates/buying-guide-9-17-2012,default,pg.html
https://www.inlineskates.com/Buying-Guide-for-Kids-Inline-Skates/buying-guide-11-13-2012,default,pg.html
https://www.inlineskates.com/Buying-Guide-for-Inline-Hockey-Skates/buying-guide-9-3-2012,default,pg.html
https://www.inlineskates.com/Buying-Guide-for-Aggressive-Skates/buying-guide-7-19-2012,default,pg.htmlThe new URLS would look like this which would be a great improvement
https://www.inlineskates.com/Learn/Buying-Guide-for-Inline-Skates,default,pg.html
https://www.inlineskates.com/Learn/Buying-Guide-for-Kids-Inline-Skates,default,pg.html
https://www.inlineskates.com/Learn/Buying-Guide-for-Inline-Hockey-Skates,default,pg.html
https://www.inlineskates.com/Learn/Buying-Guide-for-Aggressive-Skates,default,pg.htmlMy worry is that we do rank fairly well organically for some of the content and don't want to anger the google machine.
The way I would be doing the process would be to edit the URLS to the new layout, then do the redirect for them and push live.
Is there a great SEO risk to doing this?
Is there a way to do a mass "Fetch as googlebot" to reindex these if I do say 50 a day? I only see the ability to do 1 URL at a time in the webmaster backend.
Is there anything else I am missing?I believe this change would overall be good in the long run but do not want to take a huge hit initially by doing something incorrectly. This would be done on 5- to a couple hundred links across various sites I manage.
Thanks in advance,
Chris Gorski -
Hey K.
Happy you found value in everyone's response here.
For your URL, I see your structure resemble a blog structure with your subfolder being "learn" instead of "blog". So if it is, your final URL that you described is fine (learn/buying-guide-for-inline-skates).
As a counterpoint, if you had this content as a hub page (some form of content pillar/topic cluster) for example, it would've been a good possibility to just change the URL structure since you have many Buying Guides. Different types of content, different ways to put it on your site.
Like so: /buying-guide/inline-skatesAt the end of the day, the structure needs to be logical and reflective of where your content is. I think you got it right anyway.
For the execution part;
- I would not recommend using the "crawl as Googlebot" function in search console. It would be way too time-consuming for you, and it is not really designed for that kind of work.
- Instead, update your sitemap with the final URLs and send it again via SearchConsole / Bing Webmastertools.
- Also, don't forget to go ahead and change the internal links pointing to the old URL to point directly to the new ones or else you'll just have a bunch of 301's crawled by Google. Make it seamless.
- Monitor. Monitor. Monitor.
Hope that helped!
-
Thanks all the for responses and I am taking to heart all your suggestions. I will make all the URLs lowercase as this is something I didn't take into account.
Due to site hosting platform limitation for the URL structure I need to have a value in the bold areas below:
https://www.inlineskates.com/learn/buying-guide-for-inline-skates,default,pg.html
The first value "learn" doesn't have to be unique but the 2nd "learn/buying-guide-for-inline-skates" is the driver for the URL and must be unique, so the short URL like https://www.inlineskates.com/buying-guide/Inline-Skates,default,pg.html wouldn't work since I would only be able to use it 1x for inline skates and have several content pieces that I would need to add that to, size guides, charts, etc.
My main concern is about the process of doing the redirects, say I do 50 in one day, what is my next step? Is there a way to run the fetch as googlebot to a handful of pages as I only see the ability to add 1 URL at a time in the webmaster backend. If I go ahead and do this I just want to do it in the smartest way possible.
Thanks, Chris
-
And to add to that, Moz has a great resource on how to write them (emphasis mine);
-
Keeping URLs as simple, relevant, compelling, and accurate as possible is key to getting both your users and search engines to understand them.
-
URLs should be definitive but concise.
-
When necessary for readability, use hyphens to separate words. URLs should not use underscores, spaces, or any other characters to separate words.
-
Use lowercase letters. In some cases, uppercase letters can cause issues with duplicate pages.
-
Avoid the use of URL parameters, if possible, as they can create issues with tracking and duplicate content. If parameters need to be used (UTM codes, e.g.), use them sparingly.
Source - https://moz.com/learn/seo/url
-
-
The new URLs seem pretty verbose.
If this was my site, I would consider...
https://www.inlineskates.com/Buying-Guide/Inline-Skates/
https://www.inlineskates.com/Buying-Guide/Kids-Inline-Skates/
https://www.inlineskates.com/Buying-Guide/Inline-Hockey-Skates/
https://www.inlineskates.com/Buying-Guide/Aggressive-Skates/My breadcrumbs would look similar....
Home >> Buying Guide >> Inline Skates
-
I've heard recent claims that you no longer lose link juice when 301ing, but I've also had personal experience with taking a hit to organic traffic when doing this type of major URL restructuring. You're right, in the long run, it does pay off, but you're also right that there may be a ding in the short term.
While you're in there, I would highly, highly recommend taking the additional step of removing ",default,pg" from your URLs. There are two problems with that string - one is it contains commas, which can be problematic. Anything other than plain text and hyphens is generally discouraged in URLs. The other problem is that "default pg" part is watering down your URL and does not contribute to making the links short and easy to type and understand for humans. Since you are undertaking such a massive restructuring, now is the time to make these additional tweaks - as it's less painful to do All The Things at once and take one hit, than it would be to take a hit now and another hit later when you tackle that separately. I personally also always use only lowercase characters in my URLs as it's easier to type and depending on the server OS improper capitalization may make it so they don't reach the right page, but that is more of a UX preference than an SEO enhancement.
Another couple of tips: I find it extremely helpful to map everything out on a spreadsheet in this type of migration. Helps me make sure I have the old and the new mapped out, and also helps to have a checklist to go through systematically. I also tend to group related pages in batches, as presumably related pages will be linked to each other, and once Google comes crawling back over one page it will see the new URLs and crawl those too, faster than it would crawl unrelated pages. Finally, use a HTTP header check tool (google that and you'll find several) after you add the redirects just to make 100% certain you've set everything up correctly. And where possible, mass-redirect in a single .htaccess rule rather than one by one.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Beta Site Removal best practices
Hi everyone.
Intermediate & Advanced SEO | | bgvsiteadmin
We are doing a CMS migration and site redesign with some structural changes. Our temporarily Beta site (one of the staging environments and the only one that is not behind firewall) started appearing in search. Site got indexed before we added robots.txt due to dev error (at that time all pages were index,follow due to nature of beta site, it is a final stage that mirrors live site) As an remedy, we implemented robots.txt for beta version as : User-Agent: *
Disallow: / Removed beta form search for 90 days. Also, changed all pages to no index/no follow . Those blockers will be changed once code for beta get pushed into production. However, We already have all links redirected (301) from old site to new one. this will go in effect once migration starts (we will go live with completely redesigned site that is now in beta, in few days). After that, beta will be deleted completely and become 404 or 410. So the question is, should we delete beta site and simple make 404/410 without any redirects (site as is existed for only few days ). What is best thing to do, we don't want to hurt our SEO equity. Please let me know if you need more clarification. Thank you!0 -
Getting SEO Juice back after Redirect
Hi, On my website, many product pages were redirected over time to its product category, due to the product being unavailable. I understand with a 301 redirect, the final URL would have lost about 15% of the link juice. However - if after some time (e.g. 2 months, or 1 year) I remove the redirection - is the original page going to have any SEO juice, or did it already lose all of it? Thanks,
Intermediate & Advanced SEO | | viatrading10 -
Best practice for retiring old product pages
We’re a software company. Would someone be able to help me with a basic process for retiring old product pages and re-directing the SEO value to new pages. We are retiring some old products to focus on new products. The new software has much similar functionality to the old software, but has more features. How can we ensure that the new pages get the best start in life? Also, what is the best way of doing this for users? Our plan currently is to: Leave the old pages up initially with a message to the user that the old software has been retired. There will also be a message explaining that the user might be interested in one of our new products and a link to the new pages. When traffic to these pages reduces, then we will delete these pages and re-direct them to the homepage. Has anyone got any recommendations for how we could approach this differently? One idea that I’m considering is to immediately re-direct the old product pages to the new pages. I was wondering if we could then provide a message to the user explaining that the old product has been retired but that the new improved product is available. I’d also be interested in pointing the re-directs to the new product pages that are most relevant rather than the homepage, so that they get the value of the old links. I’ve found in the past that old retirement pages for products can outrank the new pages as until you 301 them then all the links and authority flow to these pages. Any help would be very much appreciated 🙂
Intermediate & Advanced SEO | | RG_SEO0 -
Link-building best SEO practice (one-off VS periodic blogging)
Hi all, Generally, what would be best when building a website's ranking through link building? Having the same links from the same bloggers or receiving new links from different bloggers every time? A lot of the SEO services offer 4-8 blog backlinks per month. Would it be best if these links came from different sources every time or most from the same sources each month? I know there's a lot of factors but I hope this question is clear. Happy holidays and thank you for your insightful feedback. Carlos
Intermediate & Advanced SEO | | 90miLLA0 -
Best practices for a local business move
My client is a chiropractic office that will be moving to the next town over in about 3 months. What are people's best practices on how to best accomplish this SEO-wise so as to not lose too much in terms of rankings, current organic traffic and citation listings?
Intermediate & Advanced SEO | | easystreetint0 -
New Site Structure and 404s. Should I redirect everything?
Hi fellow Mozzers, I've recently re-released a site and also took the opportunity to change/clean up the URL structure. As a result of this Google is starting to report many 404s such as below; blog/tag/get-fit/ blog/tag/top-presents/ Most of these 404 errors are from tag or category pages which simply don't exist any more (because they were unnecessary, crap or irrelevant). Although there's also a few posts I've removed. My question is whether it's worth redirecting all these tags and pages to the root directory of the site's new blog (as there isn't really a new page which is similar or appropriate) or just to leave them as 404 errors. Bearing in mind; They don't really rank for anything There's little if any links pointing to these pages Thanks.
Intermediate & Advanced SEO | | PeterAlexLeigh0 -
URL Shorteners. Are they SEO Friendly?
Do URL shortener services like bit.ly act as 301 redirects? I was thinking about utilizing one for longer query based URLs and didn't want to risk losing link juice. Thanks for the insight! Regards - Kyle
Intermediate & Advanced SEO | | kchandler0