Best practices for temporary articles
-
Hello,
I would like to have expert inputs about the best way to manage temporary content?
In my case, I've a page (ex : mydomain.com/agenda) where I have listing of temporary article, with a lifetime of 1 month to 6 months for some of them.
My articles also have a specific url like for ex : mydomain.com/agenda/12-02-2011/thenameofmyarticle/
As you can guess, I got hundreds of 404
I'm already using canonical tag, should I use a in the listing page? I'm a bit lost here..
-
Thanks you Egol
-
thanks Richard.
I'm going to try this.
-
Thanks Aran!
-
{script to test page URL}
$location = "http://www.YourSite.com/";
header("HTTP/1.1 301 Moved Permanently");
header("Location: {$location}");
exit;
}
-
We have temporary content and evergreen content.
When a page of temporary content is created it is filed in a folder according to its "expiration date". On that date the folder is 301 redirected to an appropriate destination. However, before the redirect is done we run analytics on the folder to see if any files are pulling traffic from SERPs or links from other websites. We then try to create evergreen content on the same topic that will capture that traffic and redirect the specific files to the new evergreen content.
-
It seems so unnatural to want to actually remove content when we spend so long striving to create awesome content!
-
You can use the meta robots tags as you mentioned in your question, this will prevent search engines indexing the pages, unfortunately we need to tackle the human side of the issue,if anyone links to the article, then eventually the link will result in a 404 page.
There is nothing wrong with a 404 page, they serve an imporant purpose. Since your articles are not around very long and not being indexed by search engines I see no reason to simply leave the 404 in place.
Ensure you have a custom 404 which is an imformative and helpful resource rather than a simple 404 Page not found message. use the 404 to direct the visitor to a category level page which is related to the topic of the article. Offer a simple list of links to various parts of the site that may be of interest.
Check out the SEOmoz articles
www.seomoz.org/blog/personalizing-your-404-error-pages
www.seomoz.org/blog/are-404-pages-always-bad-for-seo
Hope this helps.
-
I agree with Aran, setup an archive system that keeps the articles under the same URL but does not show them live on the website.
Alternatively you could setup a dumping "archive" folder where you drop all old articles in and use this link as your rel canonical link
-
Hello Arcanis,
Yes we have a destination URL for these contents, I just don't know how I can manage it when it disappears...
-
Hello Aran,
Thanks for your answers!
Unfortunately no, since the content is very "dated" (ex : 3 days music festival, etc.), we don't keep archive of this kind of content.
-
If you are using canonical tag, what is the context for that tag? do you already have a destination URL for these temporary articles?
-
Would it be possible to 'Archive' articles after the 1-6month period ?
Archive could just be a database flag that keeps the articles from appearing in Article index thus keeping the same url, but not clogging up main site with hundreds of links to expired articles?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Website blog is hacked. Whats the best practice to remove bad urls
Hello So our site was hacked which created a few thousand spam URLs on our domain. We fixed the issue and changed all the spam urls now return 404. Google index shows a couple of thousand bad URLs. My question is- What's the fastest way to remove the URLs from google index. I created a site map with sof the bad urls and submitted to Google. I am hoping google will index them as they are in the sitemap and remove from the index, as they return 404. Any tools to get a full list of google index? ( search console downloads are limited to 1000 urls). A Moz site crawl gives larger list which includes URLs not in Google index too. Looking for a tool that can download results from a site: search. Any way to remove the URLs from the index in bulk? Removing them one by one will take forever. Any help or insight would be very appreciated.
Technical SEO | | ajiabs1 -
I am looking for best way to block a domain from getting indexed ?
We have a website http://www.example.co.uk/ which leads to another domain (https://online.example.co.uk/) when a user clicks,in this case let us assume it to be Apply now button on my website page. We are getting meta data issues in crawler errors from this (https://online.example.co.uk/) domain as we are not targeting any meta content on this particular domain. So we are looking to block this domain from getting indexed to clear this errors & does this effect SERP's of this domain (**https://online.example.co.uk/) **if we use no index tag on this domain.
Technical SEO | | Prasadgotteti0 -
Changed all product titles, lost google schema markup in listings. Temporary?
We changed all of our product titles to be way shorter and less keyword stuffed last week. Short of dropping a few places in rank for most keywords (we assume temporarily) that all went fine. What we didn't expect was to loose all the schema data in our google listings from product pages. Price, and review stars are missing. Has anyone seen this before?
Technical SEO | | monkeyevil0 -
How can you best use additional domains with important keywords
Currently I have a corporate website that is ranking all right. However, I have some additional domains containing import search terms that I would like to use to get higher rankings for the corporate website, or allow these domains to generate more traffic for the corporate website. What are best practice in using these domains with keyword terms, to make most use of them, for ideally both ranking as well as generating additional traffic. All input is highly appreciated.
Technical SEO | | moojoo0 -
Internal links best practices
In looking at the inbound links to a client’s Home page, I see that the link from each page of the website back to the Home page is an image, and the ALT text is “Home.” I have a few questions about this, and would appreciate help understanding best practices: --Does it matter that the link back to the Home page is an image (presumably the client’s logo)? -- If we keep the image link, wouldn’t it be better to use “client’s company name” as ALT text rather than “Home”? --Should I recommend using an HTML link back to the Home page, and using the company name as anchor text? (I don't think it's relevant, but the site is built in Drupal.) Thanks!
Technical SEO | | jrae0 -
Deleting Subdomain - 301 to Homepage Best Option?
We have a subdomain with lots of content that we think Google may consider thin, so we're thinking of removing it to improve our SEO. Is the best way to do this to simply remove the directory and then 301 everything to our homepage? Basically the subdomain consists of product images with links to retailers where they can be purchased. We've basically used it to refer our Facebook fans to when they like a product they've seen on our Facebook page, so the subdomain was not meant to rank for SEO purposes. However, it is integrated with our main webpage, and it is possible that it is hurting our SEO efforts. The subdomain is photos.yournextshoes.com and the main domain www.yournextshoes.com
Technical SEO | | Jantaro0 -
Best Way To Clean Up Unruly SubDomain?
Hi, I have several subdomains that present no real SEO value, but are being indexed. They don't earn any backlinks either. What's the best way of cleaning them up? I was thinking the following: 1. Verify them all in Webmaster Tools. 2. Remove all URLs from the index via the Removal Tool in WMT 3. Add site-wide no-index, follow directive. Also, to remove the URLs in WMT, you usually have to block the URLs via /robots.txt. If I'd like to keep Google crawling through the subdomains and remove their URLs, is there a way to do so?
Technical SEO | | RocketZando0 -
What is the best practice to handle duplicate content?
I have several large sections that SEOMOZ is indicating has duplicate content, even though the content is not identical. For example: Leather Passport Section - Leather Passports - Black - Leather Passposts - Blue - Leather Passports - Tan - Etc. Each of the items has good content, but it is identical, since they are the same products. What is the best practice here: 1. Have only one product with a drop down (fear is that this is not best for the customer) 2. Make up content to have them sound different? 3. Put a do-no-follow on the passport section? 4. Use a rel canonical even though the sections are technically not identical? Thanks!
Technical SEO | | trophycentraltrophiesandawards0