How to safely reduce the number of 301 redirects / should we be adding so many?
-
Hi All,
We lost a lot of good rankings over the weekend with no obvious cause. Our top keyword went from p3 to p12, for example.
Site speed is pretty bad (slower than 92% of sites!) but it has always been pretty bad. I'm on to the dev team to try and crunch this (beyond image optimisation) but I know that something I can effect is the number of 301 redirects we have in place.
We have hundreds of 301s because we've been, perhaps incorrectly, adding one every time we find a new crawl error in GWT and it isn't because of a broken link on our site or on an external site where we can't track down the webmaster to fix the link. Is this bad practice, and should we just ignore 404s caused by external broken URLs?
If we wanted to reduce these numbers, should we think about removing ones that are only in place due to external broken URLs?
Any other tips for safely reducing the number of 301s?
Thanks, all!
Chris
-
Really helpful. Thanks very much, Sha. Much appreciated!
-
Hi again Chris,
OK, well that makes it interesting!
First, if the list of 301's in your .htaccess file numbers in the hundreds, then there is definitely cause for concern about the effect it might be having on your load times.
The .htaccess file is read from top to bottom until a rule is matched. The first one that is matched will be used and no rule after that will be checked. Obviously, if there are hundreds of rules to check, this can cause processing bottlenecks.
Now there are some things you might be able to do to alleviate the problem IF you have the right conditions.
-
If your rebuild involved the relocation of an entire directory or directories which still contain the same pages as before, you can write a 301 Redirect for all pages in a directory with a single line of code
-
If there is a database behind your site and you have some means of matching pages from the old URL to the new (for example, is there a unique product ID for each page?) then you can use database lookups to write the redirects on the fly. This will eliminate the processing bottlenecks, but can only work if you have a reliable means of matching the pages.
-
If you have only one competition or offer at a time, then you could use database lookups to match any page with say, "competition" in the URL and 301 it to the current competition page (and the same with "offer").
-
For unknown links from external sites you can use a "catch-all" 301 to catch any other page that returns a 404 and send it to a single page (you put this rule at the very end of your .htaccess so it is the last to be matched). You could send these all to home, a generally relevant landing or category page, or a specially designed 404 page. The most important thing is that the page you redirect them to is as relevant as possible, or provides options that may keep the visitor on your site (search, menus etc). I prefer not to send these "random" 404's to home as they are likely to bring an increase in bounce rate. Since bounce is now acknowledged as a ranking factor, I prefer to keep any traffic with a higher probability of bouncing away from pages with the highest Page Authority (PA).
Hope that helps,
Sha
-
-
Yeah, that seems to be the consensus - thanks, Aaron!
Back to the drawing board diagnosing this drop - avg rank over our top 200 keywords has gone from 80-something to 120-something. Not good at all!
Chris
-
Hi Sha,
A combination of things, really. We rebuilt and got rid of a load of legacy pages so there's a few for that reason. We also have a lot of time-limited pages like competitions and special offers that require specific landing pages. The majority are caused by broken links on external sites though - links that never existed, or that have been crawled and added to an autogenerated page incorrectly. The usual nonsense.
So, in brief, there isn't really one cause!
Many thanks for your reply.
Chris
-
In general, 301 aren't bad unless you have a whole string of them. For instance if a 301 redirected to another 301, then another, etc... If 301's are shallow they usually do not present a problem.
-
Hi BaseKit,
What is the reason for there being so many 404's?
Did you move your site or rebuild your site structure? Do you have a lot of pages that are removed after a short time?
The answer to these questions will help to know what is the best approach for your situation.
Sha
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Redirect 301 to homepage
Hello there, I have some low quality pages in my site, can I redirect 301 them to my homepage? My website is: https://idanito.com idanito.com
Technical SEO | | dannybaldwin0 -
How to resolve warning of pages with redirect chain when its your http:// to https://www.
how do I write a 301 redirect in the htaccess file so that http:// goes straight to https://www. Moz replyEli profileHey there!Thanks for reaching out to us!
Technical SEO | | VelocityWebsites0 -
Proper 301 redirect code for http to https
I see lots of suggestions on the web for forwarding http to https. I've got several existing sites that want to take advantage of the SSL boost for SEO (however slight) and I don't want to lose SEO placements in the process. I can force all pages to be viewed through the SSL - that's no problem. But for SEO reasons, do I need to do a 301 redirect line of code for every page in the site to the new "https" version? Or is there a way to catch all with one line of code that Google, etc. will recognize & honor?
Technical SEO | | wcksmith10 -
Should existing canonical tags be removed where a 301 redirect is the preferred option?
Hi, I'm working on a site that is currently using canonical tags to deal with www and non-www variations. My recommendation is to setup 301 redirects to deal with this issue instead. However, is it ok to leave the existing canonical tags in place alongside the new 301 redirects or should they be removed? My thoughts are that this is not a canonical issue and therefore they should be removed? If 301 redirects are not possible it would be better have them that nothing at all but I don't think we need both, right? Any feedback much appreciated!
Technical SEO | | MVIreland0 -
301 Redirects on Large Real Estate Website
Hi guys,We are about to move over to a new website and need advice on handling the 301 redirects.We have a large real estate website with around 12,000 pages, a lot of these are properties (about 10,000)On our old website, the url structure for each property is as follows -domainname.com/property/view?property=14863on our new site, the url structure isdomainname.com/properties/view/6137The property ID number is always different from old site to new. The way we see it, we have two options. a.) a manual redirect of each and every property url. A very very long jobb.) a folder level redirect, so redirect the 'property' folder on the old site into the 'properties' folder on new. The con with this one is we are not sure if this is the best route to take, if it is how we would go about it?Some advice would be really appreciated guys. I know there are some hyper intelligent SEO's in here and we need to make sure we handle this right!Many thanks in advance.Mark
Technical SEO | | Nextman0 -
Does any one have experience with SEO and .NET using 301 redirects?
A while ago I altered some of the URL's of my website. Google now thinks that I have two duplicate pages (duplicate content), I have asked my third party web developers (Who use .NET and a custom built CMS system) to simply 301 redirect the old URL to the other. However, my web developers say the following: "Solving the problems by 301 permanent re directs are out of the question as this would create infinite loops. Likely to bring down our server." They also wont do a canonical, as they say there is only one page (but two URLs) Firstly, has any one heard of this before and do they think this is true? Also, does anyone have an alternative method of getting rid of the old URL? Any thoughts would be much appreciated.
Technical SEO | | CoGri0 -
301 redirect from root to /index.aspx
I have taken over the SEO for www.domain.net. The way i've inherited the setup is that www.domain.net is 301 redirected to www.domain.net/index.aspx Looking at top pages and linking root domains in Opensiteexplorer I can see that www.domain.net/index.aspx has 1,006 linking root domains www.domain.net has 806 linking root domains. I assume that www.domain.net is passing the value of it's 806 domain links to www.domain.net/index.aspx via the 301 redirect and because of this would expect www.domain.net/index.aspx to be the strongest page on the site and be the url that ranks in the listings for many relevant searches. It appears however that www.domain.net is what is shown in listings and not www.domain.net/index.aspx ?? Can anyone explain why this might be?? If I do a site: search in Google then www.domain.net is indexed and not www.domain.net/index.aspx ??
Technical SEO | | QubaSEO0 -
I have a site that has both http:// and https:// versions indexed, e.g. https://www.homepage.com/ and http://www.homepage.com/. How do I de-index the https// versions without losing the link juice that is going to the https://homepage.com/ pages?
I can't 301 https// to http:// since there are some form pages that need to be https:// The site has 20,000 + pages so individually 301ing each page would be a nightmare. Any suggestions would be greatly appreciated.
Technical SEO | | fthead90