Best Practice to Remove a Blog
-
Note: Re-posting since I accidentally marked as answered
Hi,
I have a blog that has thousands of URL, the blog is a part of my site.
I would like to obsolete the blog, I think the best choices are
1. 404 Them: Problem is a large number of 404's. I know this is Ok, but makes me hesitant.
2. meta tag no follow no index. This would be great, but the question is they are already indexed.
Thoughts?
Thanks
PS
A 301 redirect to the main page would be flagged as a soft 404
-
Subdomains are treated slightly differently by Google. Essentially they are seen as less connected to the rest of your content than a subfolder.
Take wordpress.com as an example:
- surferdude.wordpress.com has little relation to www.wordpress.com
- surferdude.wordpress.com has little relation to skaterguy.wordpress.com
- surferdude.wordpress.com has lots in common with surferdude.wordpress.com/surfboards/***
In the same regard, www.yourdomain.com/blog is more correlated with www.yourdomain.com than blog.yourdomain.com would be.
By using www.yourdomain.com/blog instead of a subdomain, you build more value to your www subdomain, everytime you post blog content or get links to your blog. This has more value to the rest of the www content on your site.
-
I agree also. Thank you
As far as subdomain or subfolder, I see no difference. Can you explain Kane?
-
Agree with Kane. If you're going to be building a blog elsewhere then just setup a 301 redirect to that.
-
In that case, it doesn't sound like there are any blog posts that get frequent traffic from referrals? If that's the case, everything should get a broad 301 redirect to the new blog page. This can typically be done in one redirect depending on your URL structure, so you don't have to do each and every URL.
On the topic of subdomains, subfolders are typically a better choice for SEO purposes.
-
The blog has little value, with almost no user traffic.
It will be redesigned in a subdomain on the site.
I am only concerned with crawlers/google crawlers etc..and being penalized for tons of missing pages by 404'ing
There is nothing linking to the blog
-
A few other questions for you first:
- Why on earth are you getting rid of everything?
- Are you going to replace that content with new content - either now or eventually?
- Is there any other content on your site that is relevant to the articles?
A few broad answers that I can say without hesitation:
- No, absolutely do not leave a bunch of 404s. IMO, everything should 301 somewhere. Sending people to relevant content is best, but sending them all to the homepage or a landing page that says "sorry but we deleted our blog" is better than a 404.
- No, "noindex/nofollow" is not worthwhile. If you want to keep the content and deindex it, choose "noindex/follow." At least then you keep some of the value of the pages (they can continue spreading some of their value to other pages on the site).
-
Hiya,
Without knowing a little more about your site and the blog here are some things I would consider:
I'm going to assume that you're trying to decide what to do with the blog while still retaining the maximum benefits for the overall seo of your site.
You say that the blog has thousands of URLs. What you need to do is determine how many sites are linking to your blog content. (You can do this using Open Site Explorer or look in Google webmaster tools or Google Analyrics to see who is reffering traffic.
The first question I would ask is whether you need to remove the content at all? Would it be possible just to put up a banner on top of the existing pages to say that the blog is no longer active.
How many search visitors does the blog get? If the blog posting are getting visitors, then you need to ask yourself if you're happy to give these up.
Would anyone else be interested in taking over the blog?
If you decide to remove you content:
Put 301 redirects to direct traffic to you main site. You'll preserve some of the value of your inbound links.
Do your blog pages relate to specific content on the main site that may be of interest to the visitor? If you can determine specific pages that are strongly related to the removed pages then link to those.
I wouldn't just remove the pages and respond with a 404 error. You'll lose any value from the links to those pages.
Hope this helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Question on Pagination - /blog/ vs /blog/?page=1
Question on Pagination Because we could have /blog/ or /blog/?page=1 as page one would this be the correct way to markup the difference between these two URL? The first page of a sequence could start with either one of these URLs. Clarity around what to do on this first page would be helpful. Example… Would this be the correct way to do this as these two URLs would have the exact content? Internal links would likely link to /blog/ so signal could be muddy. URL: https://www.somedomain.com/blog/
Technical SEO | | jorgensoncompanies
<link rel="canonical" href="https://www.somedomain.com/blog/?page=1"> URL: https://www.somedomain.com/blog/?page=1
<link rel="canonical" href="https://www.somedomain.com/blog/?page=1"> Google is now saying to just use the canonical to the correct paginated URL with page number. You can read that here:
https://developers.google.com/search/docs/advanced/ecommerce/pagination-and-incremental-page-loading But they do not clarify what to do on /blog/?page=1 vs /blog/ as they are the exact same thing. Thanks for your help.0 -
Product page Canonicalization best practice
I'm getting duplicate content errors in GWT for product list pages that look like this: -www.example.com/category-page/product
Technical SEO | | IceIcebaby
-www.example.com/category-page/product/?p=2 The "p=2" example already has a rel=canonical in place, " Shouldn't the non-canonical pages be using the canonical attribute for the first page rather than the additional product pages? Thanks!0 -
Best Way to Break Down Paginated Content?
(Sorry for my english) I have lots of user reviews on my website and in some cases, there are more than a thousand reviews for a single product/service. I am looking for the best way to break down these reviews in several sub-pages. Here are the options I thought of: 1. Break down reviews into multiple pages / URL http://www.mysite.com/blue-widget-review-page1
Technical SEO | | sbrault74
http://www.mysite.com/blue-widget-review-page2
etc... In this case, each page would be indexed by search engines. Pros: all the reviews are getting indexed Cons: It will be harder to rank for "blue widget review" as their will be many similar pages 2. Break down reviews into multiple pages / URL with noindex + canonical tag http://www.mysite.com/blue-widget-review-page1
http://www.mysite.com/blue-widget-review-page2
etc... In this case, each page would be set to noindex and the canonical tag would point to the first review page. Pros: only one URL can potentially rank for "blue widget review" Cons: Subpages are not indexed 3. Load all the reviews into one page and handle pagination using Javascript reviews, reviews, reviews
more reviews, more reviews, more reviews
etc... Each page would be loaded in a different which would be shown or hidden using Javascript when browsing through the pages. Could that be considered as cloaking?!? Pros: all the reviews are getting indexed Cons: large page size (kb) - maybe too large for search engines? 4. Load only the first page and load sub-pages dynamically using AJAX Display only the first review page on initial load. I would use AJAX to load additional reviews into the . It would be similar to some blog commenting systems where you have to click on "Load more comments" to see all the comments. Pros: Fast initial loading time + faster loading time for subpages = better user experience Cons: Only the first review page is indexed by search engines ========================================================= My main competitor who's achieving great rankings (no black hat of course) is using technique #3. What's your opinion?0 -
How to set Home page for the best effect
My head is spinning with all the confusing possibilities. Does anybody have an easy answer for setting up the home page and its canonical-ishness ie Which gives the best SEO Mojo ? \ \default.aspx \keyword\ \keyword\default.aspx Thanking you in advance for reducing the number of business migranes around the globe.
Technical SEO | | blinkybill0 -
Cloaking? Best Practices Crawling Content Behind Login Box
Hi- I'm helping out a client, who publishes sale information (fashion sales etc.) In order for the client to view the sale details (date, percentage off etc.) they need to register for the site. If I allow google bot to crawl the content, (identify the user agent) but serve up a registration light box to anyone who isn't google would this be considered cloaking? Does anyone know what the best practice for this is? Any help would be greatly appreciated. Thank you, Nopadon
Technical SEO | | nopadon0 -
Best XML Generator for Wordpress?
Hi all, Quick question - does anybody have any recommendations for the best XML sitemap plugin for Wordpress? An idea of why you like it would also be helpful. Thank you very much! Mark
Technical SEO | | markadoi840 -
How to remove .html ending with mos rewrtie
I have a page and I want to remove .html ending from urls. What should I write in the htaccess?
Technical SEO | | sesertin0 -
Best blocking solution for Google
Posting this for Dave SottimanoI Here's the scenario: You've got a set of URLs indexed by Google, and you want them out quickly Once you've managed to remove them, you want to block Googlebot from crawling them again - for whatever reason. Below is a sample of the URLs you want blocked, but you only want to block /beerbottles/ and anything past it: www.example.com/beers/brandofbeer/beerbottles/1 www.example.com/beers/brandofbeer/beerbottles/2 www.example.com/beers/brandofbeer/beerbottles/3 etc.. To remove the pages from the index should you?: Add the Meta=noindex,follow tag to each URL you want de-indexed Use GWT to help remove the pages Wait for Google to crawl again If that's successful, to block Googlebot from crawling again - should you?: Add this line to Robots.txt: DISALLOW */beerbottles/ Or add this line: DISALLOW: /beerbottles/ "To add the * or not to add the *, that is the question" Thanks! Dave
Technical SEO | | goodnewscowboy0