Best course of action when removing 100's of pages from your site?
-
We had a section on our site Legal News (we are a law firm). All we did there was rehash news stories from news sites (no original content). We decided to remove the entire Legal News section and we were left with close to 800 404's. Around this same time our rankings seemed to drop. Our webmaster implemented 301's to closely related content on our blog. In about a weeks time our rankings went back up. Our webmaster informed us that we should submit each url to Google for removal, which we did. Its been about three weeks and our Not Found errors in WMT is over 800 and seems to be increasing daily. Moz's crawler says we have only 35 404's and they are from our blog not the legal news section we removed. The last thing we want is to have another rankings drop.
Is this normal? What is the best course of action when removing hundreds of pages from your site?
-
Google always takes a while to update. Check a few of the 404 errors Google is reporting and see if they are still erroneous pages. If so, you may have some pages to redirect. Of course, if the page was simply removed and you don't have anything to redirect it to, then a 404 is what is supposed to be returned. If everything looks fine, the it's just the delay of Google updating its info. It should resolve itself when they get around to updating it.
-
Hi,
It was a very bold move to drop such a significant number of pages from your site, especially if they were built up over time and attracted links. Even if the content wasn't completely original, that's not to say it didn't have some value. I think if I had made such a major change to a website and saw rankings drop, I would probably have reversed the change but then it's not clear whether that's an available option. Since I don't know the full reasoning behind the decision I'll reserve any further judgement and try to answer your question.
Returning 404s is the "right" thing to do as those pages don't exist any more, though putting 301s to very similar content is preferable to keep the benefit of any backlinks. I sense there weren't many links to worry about though as you're not very positive about the content which was deleted!
Google will hold onto pages which return 404s for some time before removing them from its index. This is to be expected as web pages can break/disappear unintentionally and so you have a grace period to "fix" any issues before losing your traffic.
The fact that Moz isn't showing any 404s shows that you aren't linking to the deleted pages because they are not being picked up by the crawl. They will drop out of WMT in a few weeks where you haven't inserted 301s to existing pages. You should also double check that they've been removed from the sitemap you submitted to Google.
Hope that helps,
George
@methodicalweb
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What is the best way to employ log-in to benefit in SEO?
Hi all, All SaaS companies have this log-in page as their top visited page in their websites and some times it helps and also hurts them. I've gone through some big SaaS companies websites and they handle the log-in page differently like on sub domain, on website page, some will directly link to their instance login without a page, etc...I wonder what is the best practice to host the log-in to make sure the more visits to log-in page don't hurt us but give us some boost. Thanks
Web Design | | vtmoz0 -
Curious why site isn't ranking, rather seems like being penalized for duplicate content but no issues via Google Webmaster...
So we have a site ThePowerBoard.com and it has some pretty impressive links pointing back to it. It is obviously optimized for the keyword "Powerboard", but in no way is it even in the top 10 pages of Google ranking. If you site:thepowerboard.com the site, and/or Google just the URL thepowerboard.com you will see that it populates in the search results. However if you quote search just the title of the home page, you will see oddly that the domain doesn't show up rather at the bottom of the results you will see where Google places "In order to show you the most relevant results, we have omitted some entries very similar to the 7 already displayed". If you click on the link below that, then the site shows up toward the bottom of those results. Is this the case of duplicate content? Also from the developer that built the site said the following: "The domain name is www.thepowerboard.com and it is on a shared server in a folder named thehoverboard.com. This has caused issues trying to ssh into the server which forces us to ssh into it via it’s ip address rather than by domain name. So I think it may also be causing your search bot indexing problem. Again, I am only speculating at this point. The folder name difference is the only thing different between this site and any other site that we have set up." (Would this be the culprit? Looking for some expert advice as it makes no sense to us why this domain isn't ranking?
Web Design | | izepper0 -
Best practices for ecommerce product categories?
I'm trying to optimise my ecommerce site's category/navigation structure so that it is: Intuitive for human users Keyword optimised, and Minimises duplicate content penalties Here is my dilemma. Let's say my site sells widgets. Some people search for widgets according to size (big widgets, medium widgets) while others search according to colour (green widgets, blue widgets). My keyword research suggests that I should target some keywords that relate to size, others that relate to colour, yet others relating to material, etc. I figured that I'd use one of these taxonomies as a category system, then set the others as filter elements. So my site's main navigation would say "Big Widgets | Medium Widgets | Small Widgets". If you click on any of them, or if you click on the "Widgets" supercategory, you'd reach a filter function allowing you to see only green widgets, or only plastic widgets, etc. So far so good - from a user perspective. The problem with this method is that Google isn't going to index my filter results. So someone Googling "green widgets" or "plastic widgets" is unlikely to find my site, even though I have plenty of green/plastic widgets that they could have filtered for. My next thought was to add some of these filter urls to my main navigation so they will be crawled. My filter mod generates urls for each filter (eg mysite.com/category?filter=k39;w24). So now I have a flashier navigation menu where clicking "Widgets" will pop out a panel allowing you to browse by size or by colour. I don't know whether users will find this helpful or redundant/confusing, but at least Google can see my filter urls. But I've run into two more problems. My filter results aren't really pages, so I can't set things like H1s, meta descriptions and so on. There's very little I can do to keyword optimise them. Further, I now have duplicate content, because the same widget can show up under multiple filter urls. And so I'm stuck here. I've thought about creating custom pages for each target keyword and manually listing products that pertain to each keyword. This will allow me to optimise the pages, but it's a lot of ongoing work (I have to update them whenever I get new stock), and I'm not sure my visitors will appreciate this - I suspect they would rather just browse/filter/search through my site than have to click through pages of manual curated content. I'd appreciate any thoughts or advice on figuring out my category and navigation system!
Web Design | | peekpeeka0 -
What causes rankings to drop while moving a site.
Hi, we recently moved a PHP based site from one web developer to another (switched hosting providers as well). Amidst the move our rankings drastically dropped and our citation and trust flow were literally cut in half as per Majestic SEO. What could have caused this sudden drop?
Web Design | | Syed_Raza0 -
How much does on-site duplicated content affect SERPs?
Hi, We've recently gotten into Moz, with our E-commerce websites, and discovered that it's crawler takes note of about 2500 pages which it thinks are the same (duplicated). We've now begun to completely rewrite every description of every product (including Meta Title/Description) so that this number may be reduced. Since this is the biggest issue Moz spots I'm wondering what the effect of fixing it will be on our position in the SERP (mainly Google). Does anybody have some stories or experience about this topic? Thanks in Advance! 🙂 Alexander
Web Design | | WebmasterAlex0 -
Hard Lessons Learned... What's yours?
So I got a whole lot of help these past two weeks and my rankings have been skyrocketing. Then I decided to start working on the on-page SEO in the lowest category of meaning, specifically on long-tail URLs. So I shortened a few of my best keyword pages so they can be fully indexed... Let's just say that I neglected to remember I had built over 2 years some 30+ PR4-6 links to these pages. Rankings for these keywords dropped from 1-2 US listings to non-existent. Lesson Learned. But I'm still smirking 🙂 What was your big lesson/mistake in the past week?
Web Design | | HMCOE0 -
What's so great about Thesis framework?
I keep hearing about how great Thesis is for SEO. But when I look at the code, it doesn't look like anything special to me -- they followed the basics (proper title, header usage, etc.), pages load quickly, and they packaged things like title and meta control with the theme itself, but none of those things seem particularly special to me. Plenty of SEO plugins give you the same control over title & meta (and the best ones go beyond what Thesis offers) and it's easy to make sure the code is clean. What am I missing?
Web Design | | EricOliver1 -
Infinite Scrolling vs. Pagination on an eCommerce Site
My company is looking at replacing our ecommerce site's paginated browsing with a Javascript infinite scroll function for when customers view internal search results--and possibly when they browse product categories also. Because our internal linking structure isn't very robust, I'm concerned that removing the pagination will make it harder to get the individual product pages to rank in the SERPs. We have over 5,000 products, and most of them are internally linked to from the browsing results pages in the category structure: e.g. Blue Widgets, Widgets Under $250, etc. I'm not too worried about removing pagination from the internal search results pages, but I'm concerned that doing the same for these category pages will result in de-linking the thousands of product pages that show up later in the browsing results and therefore won't be crawlable as internal links by the Googlebot. Does anyone have any ideas on what to do here? I'm already arguing against the infinite scroll, but we're a fairly design-driven company and any ammunition or alternatives would really help. For example, would serving a different page to the Googlebot in this case be a dangerous form of cloaking? (If the only difference is the presence of the pagination links.) Or is there any way to make rel=next and rel=prev tags work with infinite scrolling?
Web Design | | DownPour0