Best way to fix a whole bunch of 500 server errors that Google has indexed?
-
I got a notification from Google Webmaster tools saying that they've found a whole bunch of server errors. It looks like it is because an earlier version of the site I'm doing some work for had those URLs, but the new site does not.
In any case, there are now thousands of these pages in their index that error out.
If I wanted to simply remove them all from the index, which is my best option:
-
- Disallow all 1,000 or so pages in the robots.txt ?
-
- Put the meta noindex in the headers of each of those pages ?
-
- Rel canonical to a relevant page ?
-
- Redirect to a relevant page ?
-
- Wait for Google to just figure it out and remove them naturally ?
-
- Submit each URL to the GWT removal tool ?
-
- Something else ?
Thanks a lot for the help...
-
-
If you already fixed the error, then just wait for Google to figure things out on their end. Having those errors in GWT isn't going to hurt you.
-
Wouldn't you be showing 404's instead of 500's in the first place?
If the old URL's are still showing in the index, I'd reckon you'd want those 301'd to relevant pages anyways, at worst, at least a resource-heavy 404 page popping up rather than a 500.
-
4/5 with a bit of 7
What you need to do is return the correct response code (I'm guessing that is either 404 or 410) then let google reindex those URLs. That way Google knows that those urls are no longer valid. However, if those URLs have links or get traffic then you might want to 301 them.
Let's look at a couple the other options though - it is interesting.
-
This will stop google re-visiting those URLs,Therefore it will always think they are there.
-
No index confirms they are there, but tells google not to return them in results. Again this isn't correct and they will continue to return to and re-check those URLs
-
Unless the content is very close, this is unlikely to work. It is also wrong (because presumably they are not the same thing)
-
If they URLs have a common (and exclusive) directory it may be an option to submit that. It might though not be a good idea to submit lots individually - Matt Cutts has suggested this in the past.
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best Nofollow Practice Facing +500 instances of affiliate links
Hi. I have a set of affiliate links on my website. That's 50 unique links in total. They are placed in a separate folder /go/, and are all nofollowed. My website has 100 pages. When doing a site audit I get a report with 50 affiliate links pr. page x100. That's a huge report. Is this really necessary? I can ofc. choose to ignore the automated reports, but I can't help but to think that there must be an easier way to keep track of it? Like, if Link 1-50 is located in /go/, they should automatically be nofollowed, regardless of where they are located on the root/sub domain. Bonus question: The format of the link is https://exampledomain.com/go/brandname I am wondering if I should add that "/" at the end of it. If yes, why? If no, why?
Technical SEO | | Dan-Louis0 -
Removing Personal content from Google Index
Hi everyone, A user is complaining that her name is appearing in google search through our job ads site, so I removed such ads through Search Console, but the problem is not the ads anymore but our internal search results. The ads are no longer live but our searches has been indexed by google back then, We have been manually taking over 500 pages that included such name but more and more keep coming through pagination, we haven't found a pattern yet so pretty much any search result might have contained such name. We might get some legal issues here, did you guys got into anything similar before? We have just set some rules so that this doesn't happen again, but still can't find a way to deal with this one. Thanks in advance. PD: Not sure if this is the right category to fit it.
Technical SEO | | JoaoCJ0 -
Should I remove these pages from the Google index?
Hi there, Please have a look at the following URL http://www.elefant-tours.com/index.php?callback=imagerotator&gid=65&483. It's a "sitemap" generated by a Wordpress plug-in called NextGen gallery and it maps all the images that have been added to the site through this plugin, which is quite a lot in this case. I can see that these "sitemap" pages have been indexed by Google and I'm wondering whether I should remove these or not? In my opinion these are pages that a search engine would never would want to serve as a search result and pages that a visitor never would want to see. Attracting any traffic through Google images is irrelevant in this case. What is your advice? Block it or leave it indexed or something else?
Technical SEO | | Robbern0 -
Cannot work out why a bunch of urls are giving a 404 error
I have used the Crawl Diagnostic reports to greatly reduce the number of 404 errors but there is a bunch of 16 urls that were all published on the same date and have the same referrer url but I cannot see the woood for trees as to what is causing the error. **The 404 error links have the structure:**http://www.domainname.com/category/thiscategory/page/thiscategory/this-is-a-post The referrer structure is: http://www.domainname.com/category/thiscategory/page/2/ Any suggestions as to how to unravel this would be appreciated.
Technical SEO | | Niamh20 -
Google Still Taking 2 - 3 Days to Index New Posts
My website is fairly new, it was launched about 3.5 months ago. I've been publishing new content daily (except on weekends) since then. Is there any reason why new posts don't get indexed faster? All of my posts gets +1's and they're shared on G+, FB and Twitter. My website's at www.webhostinghero.com
Technical SEO | | sbrault740 -
What is the best way to remove and fight back backlink spam?
Removing low quality and spam backlinks. What is the most effective clean-up process?
Technical SEO | | matti_wilson0 -
Error msg 'Duplicate Page Content', how to fix?
Hey guys, I'm new to SEO and have the following error msg 'Duplicate Page Content'. Of course I know what it means, but my question is how do you delete the old pages that has duplicate content? I use to run my website through Joomla! but have since moved to Shopify. I see that the duplicated site content is still from the old Joomla! site and I would like to learn how to delete this content (or best practice in this situation). Any advice would be very helpful! Cheers, Peter
Technical SEO | | pjuszczynski0 -
What's the best way to deal with an entire existing site moving from http to https?
I have a client that just switched their entire site from the standard unsecure (http) to secure (https) because of over-zealous compliance issues for protecting personal information in the health care realm. They currently have the server setup to 302 redirect from the http version of a URL to the https version. My first inclination was to have them simply update that to a 301 and be done with it, but I'd prefer not to have to 301 every URL on the site. I know that putting a rel="canonical" tag on every page that refers to the http version of the URL is a best practice (http://www.google.com/support/webmasters/bin/answer.py?hl=en&answer=139394), but should I leave the 302 redirects or update them to 301's. Something seems off to me about the search engines visiting an http page, getting 301 redirected to an https page and then being told by the canonical tag that it's actually the URL they were just 301 redirected from.
Technical SEO | | JasonCooper0