What are the effects of having Multiple Redirects for pages under the same domain
-
Dear Mozers,
First of all let me wish you all a Very Happy, Prosperous, Healthy, Joyous & Successful New Year !
I'm trying to analyze one of the website's Web Hosting UK Com Ltd. and during this process I've had this question running through my mind. This project has been live since the year 2003 and since then there have be changes made to the website (obviously). There have also been new pages been added, the same way some new pages have even been over-written with changes in the url structures too.
Now, coming back to the question, if I've have a particular url structure in the past when the site was debuted and until date the structure has been changes thrice (for example) with a 301 redirect to every back dated structure, WOULD it impact the sites performance SEOwise ? And let's say that there's hundreds of such redirections under the same domain, don't you think that after a period of time we should remove the past pages/urls from the server ? That'd certainly increase the 404 (page not found) errors, but that can be taken care of.
How sensible would it be to keep redirecting the bots from one url to the other when they only visit a site for a short stipulated time?
To make it simple let me explain it with a real life scenario. Say if I was staying a place A then switched to a different location in another county say B and then to C and so on, and finally got settled at a place G. When I move from one place to another, I place a note of the next destination I'm moving to so that any courier/mail etc. can be delivered to my current whereabouts. In such a case there's a less chance that the courier would travel all the destinations to deliver the package. Similarly, when a bot visits a domain and it finds multiple redirects, don't you think that it'd loose the efficiency in crawling the site?
Ofcourse, imo. the redirects are important, BUT it should be there (in htaccess) for only a period of say 3-6 months. Once the search engine bots know about the latest pages, the past pages/redirects should be removed.
What are your opinions about this ?
-
Both answers so far get to one of the points that I was going to make, always update redirects so that there is not a chain, but I wanted to add something else. You only need redirects as long as someone is linking to those pages. You should be taking time to fix any internal references to changed URLs and contacting websites that link to the old URLs and asking them to change the URLs. That should be a part of any site URL change.
If you have only revised your URLs once, you only need redirects for 3-6 months while the search engines reindex everything. In that time, you should have changed all links to the old URLs.
In your case, I'd drop all old redirects except for the last one and see what 404s you get. Find the referring site, and contact them to change the link to your site. Once that is all done, then you can work on this latest revision to change those links.
Hope that helps!
-
It is always best to do a one to one redirect instead of a chain. As Federico said, there is some pagerank loss when doing a redirect (though the exact amount is debatable and may be neglible) and redirecting A to B to C compounds the problem. On top of that, too many redirects in a chain will lead Googlebot to stop crawling the chain. One or two is fine, three or more is not. In this older video http://youtu.be/r1lVPrYoBkA Matt Cutts started talking about redirect chains at around 2:48 and mentions that one, two and maybe three in a chain is fine. This Whiteboard Interview from 2010 with Matt Cutts http://moz.com/blog/whiteboard-interview-googles-matt-cutts-on-redirects-trust-more also states the 1 or 2 301s in a chain. So if you're redirecting A -> B -> C -> D -> E -> F... you're possibly hurting yourself. Where possible you should change the redirects so its A to F, B to F, C to F, D to F and E to F. As for removing the redirects after a certain number of months, I'd check to see how many people are still linking in with that older URL. You'd want to ask sites linking in to update to the newest URL before you 404 it and lose those links. And if you're still getting tons of direct traffic coming in on an old 301 then you might want to do some digging & research before you cut off that traffic. Odds are though after a few months you wouldn't be getting as much traffic coming through on the older URL but there is always the possibility.
-
Every time you make a redirect, 301, some of the pagerank is diluted. So following your example, from going from A to C you should redirect both A and B to C, not A -> B -> C as you double the loss.
Redirects are just fine, and in my opinion, they should say for as long as the pages being redirected still get organic traffic (backlinks, search, etc.). The moment you see no more traffic, and the links pointing to that redirected page fixed (point to the new page) you can safely remove the redirection. For as on the amount of redirects, it won't be a problem if you have lots of them, unless you do multiple redirects from A to G going from one page to the other until reaching the final, working version.
If that's not your scenario and A redirects directly to G, then you are fine. Monitor traffic on A and see if at some point you can remove the redirection, otherwise just leave is there (I personally have redirects that have been there for over 3 years as the pages are still getting organic traffic (mainly from links).
Hope that helps! And a happy new year to you too!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Sub Directories Domain & Page Crawl Depth
Hi, I just bought an old domain with good backlinks and authority, that domain was technology product formerly. So, I want to make this domain for my money site. The purpose of this website is to serve technological information like WordPress tutorial and etc (free software or drivers). And I just installed a sub directory on this domain like https://maindomain.com/subdirectory/ and this directory I made for a free software like graphics drivers download (NVIDIA or AMD). What you think with this website? Is it make sense? Wait, I just added this domain to my campaign at MOZ and the result shown my sub directory was 6 times of crawl depth. Is it good for directory or I need to move the sub directory to my main site? Thank you, hope someone answer my confuse. Best Regard, Matthew.
Intermediate & Advanced SEO | | matthewparkman0 -
How Can I Redirect an Old Domain to Our New Domain in .htaccess?
There is an old version of http://chesapeakeregional.com still floating around the web here: http://www.dev3.com.php53-24.dfw1-2.websitetestlink.com/component/content/category/20-our-services. Various iterations of this domain pop up when I do certain site:searches and for some queries as well (such as "Diagnostic Center of Chesapeake"). About 3 months ago the websitetestlink site had files and a fully functional navigation but now it mostly returns 404 or 500 errors. I'd like to redirect the site to our newer site, but don't believe I can do that in chesapeakeregional.com's .htaccess file. Is that so and would I need access to the websitetestlink .htaccess to forward the domain? Note* I (nor anyone else in our organization) has the login for the old site. The new site went live about 9 months before I arrived at the organization and I've been slowly putting the pieces together since arriving.
Intermediate & Advanced SEO | | smpomoryCRH0 -
Pages that did NOT 301 redirect to the new site
Hi, Is there a tool out there that can tell me what pages did NOT 301 redirect to the new sites? I need something rather than going into google.com and typing in site:oldsite.com to see if it's still indexed and if it's not 301 redirecting.. I'm not sure if screaming frog can do that. Thanks.
Intermediate & Advanced SEO | | ggpaul5620 -
Parking page for domain names
Hi all, I represent a hosting company which has thousands of domain names that is parked for the clients until they start using them. Currently we are presenting the client and visitors information about the situation in the top of the pages and we have placed information about all the main products in the last part of the page. You can see an example here:
Intermediate & Advanced SEO | | proisp-no
http://prodesign.no/ Would you recommend utilizing these pages in a better way than how we are doing today (SEO wise towards our own website)? We have the ability to instantly change all of these pages at once and we are also able to present different pages for every single parked domain name if we want to. Best regards,
Jon0 -
Will multiple domains for the same local business hurt its rankings?
If I have a mold removal client who has domain1.com, which is several years old, and he creates a new domain2.com with the same info (reworded), but changes over the citations to domain2 will it 1. lower domain1.com's rankings? 2. make ranking either site harder since there are now 2 sites for the same thing? 3. reduce the ability for either site to gain a higher domain authority since links can now go to either
Intermediate & Advanced SEO | | Atomicx0 -
Duplicate peices of content on multiple pages - is this a problem
I have a couple of WordPress clients with the same issue but caused in different ways: 1. The Slash WP theme which is a portfolio theme, involves setting up multiple excerpts of content that can then be added to multiple pages. So although the pages themselves are not identical, there are the same snippets of content appearing on multiple pages 2. A WP blog which has multiple categories and/or tags for each post, effectively ends up with many pages showing duplicate excerpts of content. My view has always been to noindex these pages (via Yoast), but was advised recently not to. In both these cases, even though the pages are not identical, do you think this duplicate content across multiple pages could cause an issue? All thoughts appreciated
Intermediate & Advanced SEO | | Chammy0 -
404 with a Javascript Redirect to the index page...
I have a client that is wanting me to issue a 404 on her links that are no longer valid to a custom 404, pause for 10 seconds, then rediirect to the root page (or whatever other redirect logic she wants)...to me it seems trying to game googlebot this way is a "bad idea" Can anyone confirm/deny or offer up a better suggestion?
Intermediate & Advanced SEO | | JusinDuff0 -
Multiple Domain names pointing at one website
Hello, A collegue has asked if we can buy multiple domain names which contain keywords and point them at our website. Is this good practise or will it be seen as spam? Will these domains actually get ranked? I'm sure I'm not the first person to raise this but can't seem to find any questions and answers about this. Thanks Mark
Intermediate & Advanced SEO | | markc-1971830