301 Redirect keep html files on server?
-
Hello just one quick question which came up in the discussion here: http://moz.com/community/q/take-a-good-amount-of-existing-landing-pages-offline-because-of-low-traffic-cannibalism-and-thin-content
When I do 301 redirects where I put together content from 2 pages, should I keep the page/html which redirects on the server? Or should I delete? Or does it make no difference at all?
-
yes thank you guys, that's how I thought about it as well and just wanted to clarify
-
Hi Heiko,
When you use 301 you remove the page from server. 301 redirects are done in server files like Linux/Unix's htaccess
example
redirect 301 /old-page.html http://www.adomain.com/new-page.phpFurther more, it is good practice to remove / update all your website links to this page.
In theory you are correct, if you leave the page up, it will likely never get viewed as the redirects happen at the server level. But the general rule would be to remove it.
-
Once you redirect a page, it will no longer be crawled so you can remove it from your server. Unless you want to keep it to possibly use again in the future...that's the only reason I could see for saving it. Sometimes I leave it out of sheer laziness
It doesn't matter for SEO if the content is there on your server or not because the redirect stops search bots from getting in there at all.
Hopefully that answers your question!
-
thanks shakur..
but I still think the original question is not answered completly or I missinterprete the google guidelines..
"remove it from your index" --> I can remove it via webmaster tools, but still can leave it on the server..
Personally I think it will make no difference If I remove it or not from the server, if I set the redirect, but would be happy to clarify this before.
-
Hi there,
Google provides some clear guidelines to make sure we all know how they want us to manage redirections.
301 – Search Engines: Your page is no longer available, and has permanently moved to a new page. Please remove it from your index and pass credit to the new page.
When to use 301
- As default – this is the preferred method
- Pages that are being permanently moved or replaced
- Domains that are permanently moved (acquisitions, rebranding, etc.)
- 404 pages and expired content (assuming relevant content or a page exists)
Canonical – (most) Search Engines: You have multiple versions of this page (or content), only index this version. You will keep the others available for people to see, but don't include them in your index and pass credit to your preferred page.
When to use canonical:
- When 301s can't be implemented, or take too much time
- Duplicate content but you want to keep both pages live
- Dynamic pages with multiple URLs of a single page (from sorting features, tracking options, etc.)
- Site-wide considerations like (domain/page/index.html vs. domain/page/ for the same page) can be easier with canonicals
- Cross-domain considerations where both sites are similar, but need to remain live
Hope it helps you.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Website url structure after redesign and 301 redirect chains - Looking for advice
OK, been trying to piece together what is best practice for someone I'm working with, so here goes; Website was redesigned, changed urls from url a to url b. 301's put in place. However, the new url structure is not optimal. It's an e-commerce store, and all products are put in the root folder now: www.website.com/product-name A better, more organized url structure would be: www.website.com/category/product-name I think we can all agree on that. However, I'm torn on whether it's worth changing everything again, and how to handle things in terms of redirects. The way I see things, it would result in a redirect chain, which is not great and would reduce link equity. Keeping the products in the root moving forward with a poor structure doesn't feel great either. What to do? Any thoughts on this would be much appreciated!
Technical SEO | | Tomasvdw0 -
Is there a way to set up 301 auto redirects from 404s
some of our pages under a specific website section gets deleted from another data source and we want to resolve the problem of 404s can we set up automated 301 redirects to the main page as soon as one of these pages are deleted
Technical SEO | | lina_digital2 -
Will a Robots.txt 'disallow' of a directory, keep Google from seeing 301 redirects for pages/files within the directory?
Hi- I have a client that had thousands of dynamic php pages indexed by Google that shouldn't have been. He has since blocked these php pages via robots.txt disallow. Unfortunately, many of those php pages were linked to by high quality sites mulitiple times (instead of the static urls) before he put up the php 'disallow'. If we create 301 redirects for some of these php URLs that area still showing high value backlinks and send them to the correct static URLs, will Google even see these 301 redirects and pass link value to the proper static URLs? Or will the robots.txt keep Google away and we lose all these high quality backlinks? I guess the same question applies if we use the canonical tag instead of the 301. Will the robots.txt keep Google from seeing the canonical tags on the php pages? Thanks very much, V
Technical SEO | | Voodak0 -
I have an eCommerce store with a lot of 301 redirects. Would that hurt my rankings?
I use BigCommerce, and they have a system where all the URLs are dynamically generated from the name of each product. So if I named a product "widget x y z" the url would be /widget-x-y-z/, and if I changed that to "blue widget x y z", it would change to /blue-widget-x-y-z/ and automatically redirect the old one to the new one. As a result, in 6 months, because of a lot of tweaking and experimenting, I've ended up with a hefty list of 400 redirects. Some of them are very old, and some are recent. So my question is in two parts: a) does having all of these redirects hurt my rankings? b) if so, would deleting them help?
Technical SEO | | shabbirun0 -
Loss of search engine positions after 301 redirect - what went wrong?!?
Hi Guys After adhering to the On Page optimisation suggestions given by SEOmoz, we redirected some of old urls to new ones. We set 301 redirects from the old pages to new on a page by page basis but our search engine ranking subsequently fell off the radar and lost PR. We confirmed redirection with fiddler and it shows 301 permanent redirect on every page as expected. To manage redirection using a common code logic we executed following: In Http module, using “rewrite path” we route “all old page requests” to a page called “redirect.aspx? oldpagename =[oldpagename]”. This happens at server side. In redirect.aspx we are redirecting from old page to new page using 301 permanent redirect. In the browser, when old page is requested, it will 301 redirect to new page. In hope we and others can learn from our mistakes - what did we do wrong ?!? Thanks in advance. Dave - www.paysubsonline.com
Technical SEO | | Evo0 -
How to safely reduce the number of 301 redirects / should we be adding so many?
Hi All, We lost a lot of good rankings over the weekend with no obvious cause. Our top keyword went from p3 to p12, for example. Site speed is pretty bad (slower than 92% of sites!) but it has always been pretty bad. I'm on to the dev team to try and crunch this (beyond image optimisation) but I know that something I can effect is the number of 301 redirects we have in place. We have hundreds of 301s because we've been, perhaps incorrectly, adding one every time we find a new crawl error in GWT and it isn't because of a broken link on our site or on an external site where we can't track down the webmaster to fix the link. Is this bad practice, and should we just ignore 404s caused by external broken URLs? If we wanted to reduce these numbers, should we think about removing ones that are only in place due to external broken URLs? Any other tips for safely reducing the number of 301s? Thanks, all! Chris
Technical SEO | | BaseKit0 -
301 Redirects
Hello, I have a problem with my website. I have a page on my website http://www.ensorbuilding.com/page.php/aboutus but if i type in www.ensorbuilding.com/page.php/aboutus/f8e45e9d9df6140bb5a7ff1173e8d828 or www.ensorbuilding.com/page.php/aboutus/0f0eea5e9ab0a3e8d91fad8fc0d3ce9c it still displays the about us page. Google is seeing this as duplicate content so what I would want to do is 301 redirect anything after www.ensorbuilding.com/page.php/aboutus . How could I implement a 301 redirect in this way?
Technical SEO | | danielmckay70 -
What are the SEOmoz-suggested best practices for limiting the number of 301 redirects for a given site?
I've read some vague warnings of potential problems with having a long list of 301 redirects within an htaccess file. If this is a problem, could you provide any guidance on how much is too much? And if there is a problem associated with this, what is that problem exactly?
Technical SEO | | roush0