How to Destroy Old 404 Pages
-
Hello Mozzers,
So I just purchased a new domain and to my surprise it has a domain authority of 13 right out of the box (what luck!). I needed to investigate. To make a long story short the domain used to be home to a music blog that had hundreds of pages which of course are all missing now. I have about 400 pages on my hands that are resulting in a 404. How or what is the best method for eliminating these pages.
Does deleting the Crawl Errors in Google Webmaster Tools do anything?
Thanks
-
What a thorough response! I'm in the Option B scenario. The old content has nothing to do with my site so I don't need to redirect the old URLs. I will just wait out Google crawling those 404s.
Thanks!
-
You have a few options here. Option A is if you are going to build a site that will have similar topic based content as the old one and you want to use a larger portion of that domain authority from the old site to the new.
-
Pull those 404 errors from GWT in a spreadsheet. This gives you a corpus of links to work with.
-
Go into Bing WT and they have a way to browse what they have and had indexed. What is nice here is that Bing will tell you what URLs (even old 404s) have links to them.
-
Run your links through Open Site Explorer. You can then also get linking data, FB and Twitter data in addition to OSE data on the old URLs
-
If need be, run the more important dead URLs through the Wayback Machine http://archive.org/web/web.php you can now even see what the actual content was on the old URLs.
-
After doing all of this, pretty quick you should be able to see if there were any authority pages on the site that have now expired and you also know what those pages were about via the wayback machine.
-
On the authority pages, create new pages on the new site that have to do with the same topic, i.e. semantically related to the old page.
-
301 the old authority pages to the new authority pages.
-
The rest of the URLs you can just let them 404. They will continue to 404 several time until Google drops them. I would leave them in GWT as over time they should drop out as Google starts to ignore those pages, this may take a few months. You can then just check GWT for any new 404s that might show up from the new site and you need to deal with.
One thing to note on all of this. You may have to let the old sitemap 404 vs redirecting the sitemap.
http://moz.com/blog/how-to-fix-crawl-errors-in-google-webmaster-tools
"One frustrating thing that Google does is it will continually crawl old sitemaps that you have since deleted to check that the sitemap and URLs are in fact dead. If you have an old sitemap that you have removed from Webmaster Tools, and you don’t want being crawled, make sure you let that sitemap 404 and that you are not redirecting the sitemap to your current sitemap."
If you delete the 404s from GWT the next time Google spiders the old pages they will just show up again, up to you then.
Option B - if you dont care about the old pages, just let them 404 as mentioned above, but be aware of the issue with old sitemaps. You can check the Google index for old URLs in the SERPs or also if you look into GWT and look for data on your Search Traffic. Make sure that the old URLs are not showing up under your Search Queries.
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How would you handle these pages? Should they be indexed?
If a site has about 100 pages offering specific discounts for employees at various companies, for example... mysite.com/discounts/target mysite.com/discounts/kohls mysite.com/discounts/jcpenney and all these pages are nearly 100% duplicates, how would you handle them? My recommendation to my client was to use noindex, follow. These pages tend to receive backlinks from the actual companies receiving the discounts, so obviously they are valuable from a linking standpoint. But say the content is nearly identical between each page; should they be indexed? Is there any value for someone at Kohl's, for example, to be able to find this landing page in the search results? Here is a live example of what I am talking about: https://www.google.com/search?num=100&safe=active&rlz=1C1WPZB_enUS735US735&q=site%3Ahttps%3A%2F%2Fpoi8.petinsurance.com%2Fbenefits%2F&oq=site%3Ahttps%3A%2F%2Fpoi8.petinsurance.com%2Fbenefits%2F&gs_l=serp.3...7812.8453.0.8643.6.6.0.0.0.0.174.646.3j3.6.0....0...1c.1.64.serp..0.5.586...0j35i39k1j0i131k1j0i67k1j0i131i67k1j0i131i46k1j46i131k1j0i20k1j0i10i3k1.RyIhsU0Yz4E
Intermediate & Advanced SEO | | FPD_NYC0 -
I have 2 keywords I want to target, should I make one page for both keywords or two separate pages?
My team sells sailboats and pontoon boats all over the country. So while they are both boats, the target market is two different types of people... I want to make a landing page for each state so if someone types in "Pontoon Boats for sale in Michigan" or "Pontoon boats for sale in Tennessee," my website will come up. But I also want to come up if someone is searching for sailboats for sale in Michigan or Tennessee (or any other state for that matter). So my question is, should I make 1 page for each state that targets both pontoon boats and sailboats (total of 50 landing pages), or should I make two pages for each state, one targeting pontoon boats and the other sailboats (total of 100 landing pages). My team has seen success targeting each state individually for a single keyword, but have not had a situation like this come up yet.
Intermediate & Advanced SEO | | VanMaster0 -
Removing pages from index
My client is running 4 websites on ModX CMS and using the same database for all the sites. Roger has discovered that one of the sites has 2050 302 redirects pointing to the clients other sites. The Sitemap for the site in question includes 860 pages. Google Webmaster Tools has indexed 540 pages. Roger has discovered 5200 pages and a Site: query of Google reveals 7200 pages. Diving into the SERP results many of the pages indexed are pointing to the other 3 sites. I believe there is a configuration problem with the site because the other sites when crawled do not have a huge volume of redirects. My concern is how can we remove from Google's index the 2050 pages that are redirecting to the other sites via a 302 redirect?
Intermediate & Advanced SEO | | tinbum0 -
Home page not being indexed
Hi Moz crew. I have two sites (one is a client's and one is mine). They are both Wordpress sites and both are hosted on WP Engine. They have both been set up for a long time, and are "on-page" optimized. Pages from each site are indexed, but Google is not indexing the homepage for either site. Just to be clear - I can set up and work on a Wordpress site, but am not a programmer. Both seem to be fine according to my Moz dashboard. I have Webmaster tools set up for each - and as far as I can tell (definitely not an exper in webmaster tools) they are okay. I have done the obvious and checked that the the box preventing Google from crawling is not checked, and I believe I have set up the proper re-directs and canonicals.Thanks in advance! Brent
Intermediate & Advanced SEO | | EchelonSEO0 -
How do you find the source of a 404
I've recently noticed quite a sharp rise in the number of 404 errors on our site. Unsurprising really as we have just had a major upgrade of our site navigation. When looking in GA at the landing page stats the vast majority of them are from Google organic search. How do I find the source of the 404? All help appreciated. Thanks. Jon
Intermediate & Advanced SEO | | TTS_Group0 -
410 pages
Do you need to optimize a 410 page like you do for 404 pages? What does a visitor see when a page is 410 compared to a 404?
Intermediate & Advanced SEO | | WebServiceConsulting.com0 -
Why Is This Page Not Ranking?
Hi Mozzers, I can't rank (the page is nowhere on the Google grid that I can find) and I've not been able to move the needle at all on it. The page is http://www.lumber2.com/Western-Saddle-Pads-s/98.htm for keyword "western saddle pads." I'm inclined to think I'm cannabalizing the category with the products so I removed the word saddle from the majority of the product names on page. However, saddle pad or saddle pads is in the meta title for most if not all of the products. Do you think I'm cannabalizing with the product titles or is there something else going on? Thanks for any help.
Intermediate & Advanced SEO | | AWCthreads0 -
How do I best deal with pages returning 404 errors as they contain links from other sites?
I have over 750 URL's returning 404 errors. The majority of these pages have back links from sites, however the credibility of these pages from what I can see is somewhat dubious, mainly forums and sites with low DA & PA. It has been suggested placing 301 redirects from these pages, a nice easy solution, however I am concerned that we could do more harm than good to our sites credibility and link building strategy going into 2013. I don't want to redirect these pages if its going to cause a panda/penguin problem. Could I request manual removal or something of this nature? Thoughts appreciated.
Intermediate & Advanced SEO | | Towelsrus0