Best way to remove worthless/thin content?
-
I have a Wordpress site with about 3,000 pages and 1,000 of those are no value/duplicate content and drive no traffic. They are blog posts each with a single image and permalinks like example.com/post1, example.com/post2 etc.
I've started by deleting pages and 301 redirecting to relevant pages that actually have content.
-
Is deleting and 301 redirecting the best route?
-
Is 1,000 to many 301 redirects?
-
Should I just delete the pages that aren't really relevant to anything else?
-
Anything else I should know about deleting all of these pages?
Any help would be great!
-
-
As Alan said, if the pages don't have any inbound links then you're not going to lose any link equity by removing them. If the pages in question are not getting direct traffic or search traffic and don't have any links then I'd just delete them.
Otherwise a 301 redirect to the most relevant content is the best way to go.
Do a check for any internal links to these pages too. Your own links should be easy to fix. Remove them or point them at the better relevant content.
It might also be worth reviewing your 404 page. This can often be forgotten, but it can do a lot of the hard work guiding visitors to the content they're looking for.
-
It depends how you do the 301 redirects.
I have 25,000 301s and they are not in .htaccess, otherwise it would be slow.
If nobody is linking to the ones you want to delete, and search engines send no traffic to them, just delete them and don't bother to redirect.
-
You may find that the yoast SEO plugin can help you. There are options to noindex tags pages and attachment pages, etc.
-
I know that the number of 301-redirects can sometimes raise concerns about site speed. However, 1,000 is really not that many. Large e-commerce sites can have 10,000 301s, or even more. Depending on how the 301 files is handled even this large number can be handled in ways that have very little impact on site speed or page load times.
I think the number in your vase is nothing to worry about.
I do think that deleting and 301 redirects is the best way to go for anything that's remotely relevant. For those that aren't, create a custom 404 page and let them 404. Eventually they will drop out of the index. If they don't, you could still file a remove URL request in Google Webmaster Tools.
Hope this helps and good luck!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Your Opinion: Thin Content? Should we Retire this section?
Only way to explain this was to make a video. Would love everyone's input on this: https://youtu.be/TcdaOvz24Aw
Search Behavior | | HLTalk
Thank you.0 -
Strategy for 301s/404s
Hi all, We receive a feed containing new products every morning, some products are removed and this obviously ends up in Webmaster as a 404. Is it fine to leave this as a 404 or would 301'ing the link be a better option? If we leave as a 404 we end up with thousands after a couple of month which scares me and makes finding actual 404s more difficult. Essentially, if lots of pages are being removed everyday, what's the best way of tackling these? Thanks!
Search Behavior | | HB170 -
Correct approach to a business website with separate content for personal and business customers
I'm laying the groundwork for a fairly involved website. The website is for a telco that caters to both residential and B2B. I was browsing the websites of the likes of Verizon, AT&T, Sprint & T-Mobile. What I saw is that they compartmentalize almost everything - all their business pages are in a business subdoman, all their investor info is in an investor subdomain and so-on. So I'm going implement this strategy on this website update. I just want to make sure that my idea makes sense and isn't a complete cluster****. I've attached a link to the mind map. Everything with "(sub)" attached to it is a subdomain. Everything else is a page at the root level of the top domain. Most of the visitors we get to the website are residential, so instead of loading a portal at first and asking if they're there for person or business reasons, I'm considering forwarding all visitors to the top-level domain to the personal.example.com site. Is this okay or would it be better to just keep the content in the top-level rather than forwarding all traffic to a subdomain? Thank you! 1JY7DWw
Search Behavior | | CucumberGroup0 -
Has anyone found a best practice for ranking a client who has one central office location, but a large, regional service area that they want to rank for?
All the most recent local search ranking is based upon physical addresses, but what happens when someone provides services and has staff 100 miles from that central office location, but no physical address to work with?
Search Behavior | | BehrDesign0 -
Dupe Content: Canonicalize the Wordpress Tag or NoIndex?
Mozzers, Here we go. I've read multiple posts for years on taxonomy dupe content. In fact, I've read 10 articles tonight on taxonomies and categories. A little background: I am using Wordpress SEO with the Yoast plugin. **Here is the scenario: We have 560 tags - some make sense - some do not. ** What do I do? Do I not worry about it? Matt Cutts said twice that I should not stress about it, because in the worse non-spammy case, Google may just ignore the duplicate content. Matt said in the video, “I wouldn’t stress about this unless the content that you have duplicated is spammy or keyword stuffing.” (Found Via Search Engine Land - http://searchengineland.com/googles-matt-cutts-duplicate-content-wont-hurt-you-unless-it-is-spammy-167459). Do I NoIndex,Follow the Tags? Yoast and a Moz post both say I should NoIndex and Follow the Tags. From the post: "Tag, author, and date archives will all look too similar to other content. So it does not make sense to have them indexed." BUT! **The tags have been indexed for YEARS! And both articles go onto say **"if your blog has already existed for some time, and you've been indexing tags all along for example, you shouldn't just go deindexing them" (http://moz.com/blog/setup-wordpress-for-seo-success). So do I deindex tags that have been indexed for years? I checked the analytics, and in the past month, tags have brought in less than 1% of traffic, but they are bringing in traffic. Do I canonicalize the tags? Canonicalize the URL from "http://domain.com/blog/tag/addiction/" to "http://domain.com/blog/" ? And if I canonicalize, would you canonicalize to the /blog or to the base /tag? Thanks for any and all help. I just want to clarify this issue. One of the reasons is because I received a Moz Report with a TON of dupe content warning from the tags and categories.
Search Behavior | | Thriveworks-Counseling2 -
Content marketing where articles aren't high traffic
Hello, If no one is writing articles in your niche and articles are very scarce in the top 100 landing pages, what does that tell you about content and content marketing in your niche
Search Behavior | | BobGW0 -
Best way to check the number of backlinks?
I know everyone is going to say opensiteexplorer (for obvious reasons) however why is there such a diffference in numbers between the following resources: 1. Open Site Explorer - 23 ref. domains & 939 links 2. Majestic SEO - 49 ref. domains & 599 links (fresh) / 111 ref. domains & 2332 links (historic) 3. Google Webmaster Tools - 6001 I understand the difference between fresh/historic in Majestic, it's the vast difference in numbers between the three services?!
Search Behavior | | seanjohnson0 -
Decline in engagement metrics, due to nav changes vs. content changes
With improvements in our rankings, we are seeing adverse changes in our measures of engagement. My gut reaction is to believe we are attracting more unqualified traffic, thus higher bounce rates, declines in pages/visit and time on site (approx 15%, 15%, 25%, respectively). While recent improvements in navigation might have contributed to these engagement declines, do you have any suggestions how best to determine whether these declines are due to nav changes vs. due to copy/content issues? There's been no change in copy content during this period. Thanks.
Search Behavior | | ahw0