Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Does removal of internal redirects(301) help in SEO
-
I am planning to completely remove 301 redirects manually by replacing such links with actual live pages/links. So there will be no redirects internally in the website. Will this boost our SEO efforts?
Auto redirects will be there for incoming links to non-existing pages.
Thanks,
Satish
-
Hi Dirk,
You got it right. All the pages we redirected were pointing to similar pages once, so probably they should be okay as u said.
Regarding disavow; what are the metrics to decide on a link? Some links might be good looking with decent DA and might be hurting us. What's the best way to findout actual back-links dropping us down. Disavow comes with risk as there are chances we may reject good links, so it's better to make sure about the links.
Thanks,
Satish
-
Hi Satish,
Not sure if I fully understand your answer.
Google will consider a redirect as a soft 404 if you redirect pages to non-related pages. Example: if you redirect a page about "shirts" to a page about "pants" or if you redirect this page to your homepage. If the pages are similar (example "green shirts" to "shirts") it's not considered as a soft 404. I understand that you are redirecting to similar pages - so that should be ok.
If you have pages with low quality incoming links (or a mix of high/low quality links) you can still redirect them - but in case of low quality links it's probably a good idea to disavow them (using the search console) - check https://support.google.com/webmasters/answer/2648487?hl=en
Hope this helps,
Dirk
-
Hi Dirk,
Thanks for the response with a detailed answer. Very informative. We have decent number of redirects from homepage and top tier pages too. So I decided for this activity.
Regarding "auto redirects", we have redirected tens of links these days in the process of link reclamation to increase the back-links and pagerank/da. But we have significantly dropped post such redirects even though we have cross examined to make sure the incoming links are relevant to current pages we are linking. However, those external links were pointing to our pages in past. But as we deleted many pages in process of website redesign and content update, we replaced such pages. Why actually they consider such as soft 404 even we redirect non-existing pages to almost same pages with high relevancy? I think some of the links we reclaimed are kind of spammy and pushed us down. What's your idea on this. Thank you.
-Satish
-
The crawl by Googlebot will be more efficient if it can go directly to the destination page rather than having to go trough a redirection.
There is some discussion whether 3xx redirections do have an impact on page rank / page authority - Google official point of view is that it doesn't.
Redirects do however indicate have an impact on speed (check here: https://developers.google.com/speed/docs/insights/mobile: "we strongly encourage webmasters to minimize the number, and ideally eliminate redirects entirely" - context is mobile but is applicable on all redirects) - but again for most sites this won't make a huge difference on total load time.
I doubt that simply cleaning your site and removing the 301 will give a boost to your search traffic, but it just something you need to from time to time (idem for internal 4xx errors) to improve the general health of your site.
Dirk
PS I am a bit puzzled about the remark "auto redirects" - you must make sure if you redirect that you redirect to a page which is similar to the page that has disappeared. Google considers most other type of redirects as "soft 404". If the page never existed - like domain.com/kklfjklgjkldfjg - it should return 404 and not be redirect.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Using a Reverse Proxy and 301 redirect to appear Sub Domain as Sub Directory - what are the SEO Risks?
We’re in process to move WordPress blog URLs from subdomains to sub-directory. We aren’t moving blog physically, but using reverse proxy and 301 redirection to do this. Blog subdomain URL is https://blog.example.com/ and destination sub-directory URL is https://www.example.com/blog/ Our main website is e-commerce marketplace which is YMYL site. This is on Windows server. Due to technical reasons, we can’t physically move our WordPress blog to the main website. Following is our Technical Setup Setup a reverse proxy at https://www.example.com/blog/ pointing to https://blog.example.com/ Use a 301 redirection from https://blog.example.com/ to https://www.example.com/blog/ with an exception if a traffic is coming from main WWW domain then it won’t redirect. Thus, we can eliminate infinite loop. Change all absolute URLs to relative URLs on blog Change the sitemap URL from https://blog.example.com/sitemap.xml to https://www.example.com/blog/sitemap.xml and update all URLs mentioned within the sitemap. SEO Risk Evaluation We have individual GA Tracking ID and individual Google Search Console Properties for main website and blog. We will not merge them. Keep them separate as they are. Keeping this in mind, I am evaluating SEO Risks factors Right now when we receive traffic from main website to blog (or vice versa) then it is considered as referral traffic and new cookies are set for Google Analytics. What’s going to happen when its on the same domain? Which type of settings change should I do in Blog’s Google Search Console? (A). Do I need to request “Change of Address” in the Blog’s search console property? (B). Should I re-submit the sitemap? Do I need to re-submit the blog sitemap from the https://www.example.com/ Google Search Console Property? Main website is e-commerce marketplace which is YMYL website, and blog is all about content. So does that impact SEO? Will this dilute SEO link juice or impact on the main website ranking because following are the key SEO Metrices. (A). Main website’s Avg Session Duration is about 10 minutes and bounce rate is around 30% (B). Blog’s Avg Session Duration is 33 seconds and bounce rate is over 92%
Intermediate & Advanced SEO | | joshibhargav_200 -
Is there any benefit to changing 303 redirects to 301?
A year ago I moved my marketplace website from http to https. I implemented some design changes at the same time, and saw a huge drop in traffic that we have not recovered from. I've been searching for reasons for the organic traffic decline and have noticed that the redirects from http to https URLs are 303 redirects. There's little information available about 303 redirects but most articles say they don't pass link juice. Is it worth changing them to 301 redirects now? Are there risks in making such a change a year later, and is it likely to have any benefits for rankings?
Intermediate & Advanced SEO | | MAdeit0 -
301 Redirects to relative URLs not absolute a problem?
Hi we recently did a migration and a lot of content changed locations see: https://d.pr/i/RvqI81 Basically, the 301 goes to the correct location but its a relative URL (as you can see from the screenshot) rather than absolute URL. Do you think this is a high priority issue from an SEO standpoint, should we get the developer to change the redirects to absolute? Cheers.
Intermediate & Advanced SEO | | cathywix0 -
Default Wordpress 301 Redirects of JS and CSS files. Bad for SEO & How to Fix?
Hi there: We are developers with some digital marketing expertise, but a current issue has us perplexed. An outside SEO firm has asked us to clean up a large number of 301 redirects. Most of these are 'default' Wordpress behavior that relate to calling the latest version of a JS or CSS file. For instance, a JS file is called with this: https://websitexyz.com/wp-includes/js/wp-embed.min.js?ver=4.9.1 but ultimately redirects to this: https://websitexyz.com/wp-includes/js/wp-embed.min.js. We are being asked to prevent the redirect from happening by, presumably, calling the ultimate file to begin with. The issue is that, as far as we know, there's no easy way to alter WP behavior to call the ultimate file to begin with. Does anyone have any thoughts on this? Thanks.
Intermediate & Advanced SEO | | Daaveey0 -
Should we 301 redirect old events pages on a website?
We have a client that has an events category section that is filled to the brim with past events webpages. Another issue is that these old events webpages all contain duplicate meta description tags, so we are concerned that Google might be penalizing our client's website for this issue. Our client does not want to create specialized meta description tags for these old events pages. Would it be a good idea to 301 redirect these old events landing pages to the main events category page to pass off link equity & remove the duplicate meta description tag issue? This seems drastic (we even noticed that searchmarketingexpo.com is keeping their old events pages). However it seems like these old events webpages offer little value to our website visitors. Any feedback would be much appreciated.
Intermediate & Advanced SEO | | RosemaryB0 -
Best practice for removing indexed internal search pages from Google?
Hi Mozzers I know that it’s best practice to block Google from indexing internal search pages, but what’s best practice when “the damage is done”? I have a project where a substantial part of our visitors and income lands on an internal search page, because Google has indexed them (about 3 %). I would like to block Google from indexing the search pages via the meta noindex,follow tag because: Google Guidelines: “Use robots.txt to prevent crawling of search results pages or other auto-generated pages that don't add much value for users coming from search engines.” http://support.google.com/webmasters/bin/answer.py?hl=en&answer=35769 Bad user experience The search pages are (probably) stealing rankings from our real landing pages Webmaster Notification: “Googlebot found an extremely high number of URLs on your site” with links to our internal search results I want to use the meta tag to keep the link juice flowing. Do you recommend using the robots.txt instead? If yes, why? Should we just go dark on the internal search pages, or how shall we proceed with blocking them? I’m looking forward to your answer! Edit: Google have currently indexed several million of our internal search pages.
Intermediate & Advanced SEO | | HrThomsen0 -
Iframe redirect bad for SEO?
Hi, I have a website (http://www.blowingminds.de) wich I put a spreadshirt shop into via iframe. The thing is I am not sure on how the iframe effects my SEO? Can I just optimise the main domain for search? Well I want the spreadshirt shop to be found under the domain name (www.blowingminds.de) but the only real way to do it is by implementing an iframe because each spreadshirt shop has its own subdomain eg.: blowingminds.spreadshirt.de but the only real way to do it is via iframe, as they do not offer a complete domain redirect. (Or have I overseen some other way?) I hope you guys can help me on this one 🙂 Thanks in advance. Malte
Intermediate & Advanced SEO | | wellbo1 -
Any way to find which domains are 301 redirected to competitors' websites?
By looking at the work from an SEO collegue it became clear that his weak linkbuilding graph probably is not the cause for his good rankings for a pretty competitive keyword. (also no social mentions where found) I was wondering what it could be, site structure and other on page optimization factors seems to be ok and I don't think there will be exceptionally good or bad user behavior... Finally I looked at the competitors and found that they have more links, better content en better design, so I got a little stuck. The only reason I can think of is that he is doing 301 redirects (or is rel=canonical tags). Is there a way to trace these redirects back to the source in order to include this important variable in your competitor research? thnx
Intermediate & Advanced SEO | | djingel10