Migrate Old Archive Content?
-
Hi,
Our team has recently acquired several newsletter titles from a competitor.
We are currently deciding how to handle the archive content on their website which now belongs to us.
We are thinking of leaving the content on their site (so as not to suddenly remove a chunk of their website and harm them) but also replicating it on ours with a canoncial link to say our website is the original source.
The articles on their site go back as far as 2010.
Do you think it would help or hinder our site to have a lot of old archive content added to it? I'm thinking of content freshness issues.Even though the content is old some of it will still be interesting or relevant.
Or do you think the authority and extra traffic this content could bring in makes it worth migrating.
Any help gratefully received on the old content issue or the idea of using canonical links in this way.
Many Thanks
-
Thanks for all your help with this.
-
I agree 100% with Hutch and Patrick. Your best bet is to dive into whatever analytics data you have for the content. I would probably follow a rough procedure like:
- Identify content no one is looking at, is not ranking, is old/poor - start there and you can probably trim out the lowest quality stuff - remove completlely or just noindex to be more conservative
- Then find the other extreme - think 80/20 - find the obvious highest achievers and those are the ones you'd most want to maybe move over or maintain in some way. If any high achievers are getting traffic despite being old/poor - that won't last - so update them.
- The hardest to figure out is the mediocre performing stuff (moderate visits, moderate search visibility). I would probably put all the moderate content in a spreadsheet. Categorize it by topic. Figure out what can stand alone, or be consolidated. Basically you want to arrive at a situation where every piece of content you keep is, if not recent, at least still quality (quality as defined by: unique, well executed, good design, good UX, helpful or entertaining).
The content audit process mentioned by Patrick is a great way to do this analysis with data, but you can also just use some traffic and basic segmenting in analytics as an easier method.
You could also try some tools like URL Profiler, which cake make such an audit process a little easier.
That's just decided if you should keep it - when it comes to migrating I guess it depends on your ultimate vision for the company / branding.
I wouldn't try any tricky things like putting a canonical to say your site is the original source. Google probably knows this is not true, and a canonical is just a "suggestion" so there's no guarantee they will honor it. I would be more in favor of migrating it to your site, removing from the old with a 301 redirect to your site and maybe just a note on your site saying "this article originally appeared in ...." and be really transparent with the user.
-
Great answer, Hutch.
Building on that - Moz offers an extremely comprehensive content audit that goes step by step on how to evaluate your content.
No blanket answer - this will take time and research, but it will make your site so much better overall!
Good luck!
-
I think you are asking a large, loaded question. I do not think there is a "yes you should" or "no you should not" answer for your complex question.
This content is upwards of nearly half a decade old, is it still relevant? Instead of a blanket yes or no, I think you should go through all of it and see what is still valuable, depending on your industry it could be half of it, or it could be none, but you should be looking at each piece individually, not the entire site as one whole. For moving it, if the content is good I think placing it on your site (as I assume you want to consolidate) and redirecting to the new location is fine, plus if you do it as you go, you will not have a massive surge in your content, or drop in the old site but a gradual shift over a period of time.
Hope this helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Old site selected as canonical on GSC 3 years after migration?
Recently my company started consulting for a SaaS company. They're clearly the best known, most trusted company on their area of work and they have the strongest brand, best product and therefore more users than any of their competitors by a big margin. Still, 99% of their traffic comes from branded, despite having 3x more domains, better performance scores and more content. Even using tools such as SimilarWeb for comparing user satisfaction metrics, they seem to have lower bounce rates and more visits per session. Still, they rank for almost nothing that is non branded on Google (they rank extremely well for almost everything on bing and DuckDuckGo). They don't have any obvious issues with crawling or indexation - we've gone to great depths to tick off any issues that could be affecting this. My conclusion is that it's either a penalty or a bug, but GSC is not flagging any manual actions. These are the things we've identified: All the content was moved from domain1.com to domain2.com at the end of 2017. 301s were put in place, migration was confirmed on GSC. Everything was done with great care and we couldn't identify any issues with it. Some subdomains of the site, especially support, rank extremely well for all sorts of keywords, even very competitive ones but the www subdomain ranks for almost nothing on Google. The www subdomain has 1,000s of domains pointing to it while the support has only a few 100s. Google is performing delayed rendering attempts on old pages, JS and CSS particularly versions of assets that were live before the migration in 2017, including the old homepage. Again, the redirects have been in place for 3 years. Search Console frequently showing old HTML (at least a year old) in cache despite a recent crawl date and a current 301. Search Console frequently processing old HTML (at least a year old) when reporting on schema. Search Console is sometimes selecting pages from the old domain as the canonical of a URL of an existing page of the current domain, despite a long-standing 301 and the canonicals being well configured for 3 years now. Has anyone experienced anything similar in the past? We've been doing an analysis of old SEO practices, link profile, disavow... nothing points to black hat practices and at this point we're wondering if it's just Google doing a terrible job with this particular domain.
Technical SEO | | oline1230 -
Is the content on my website is garbage?
I received a mail from google webmasters, that my website is having low quality content. Website - nowwhatmoments.com
Technical SEO | | Green.landon0 -
Content Duplication - Zencart
Hi Guys !!! Based on crawler results, it shows that I have 188 duplicate content pages, out of which some are those in which I am not able to understand where the duplication is ??? The page created is unique. All the URL's are static, all titles, metat tags are unique. How do I remove this duplication !!! I am using Zencart as a platform. Thanks in advance for the help !!! 🙂
Technical SEO | | sidjain4you0 -
Duplicate title/content errors for blog archives
Hi All Would love some help, fairly new at SEO and using SEOMoz, I've looked through the forums and have just managed to confuse myself. I have a customer with a lot of duplicate page title/content errors in SEOMoz. It's an umbraco CMS and a lot of the errors appear to be blog archives and pagination. i.e. http://example.com/blog http://example.com/blog/ http://example.com/blog/?page=1 http://example.com/blog/?page=2 and then also http://example.com/blog/2011/08 http://example.com/blog/2011/08?page=1 http://example.com/blog/2011/08?page=2 http://example.com/blog/2011/08?page=3 (empty page) http://example.com/blog/2011/08?page=4 (empty page) This continues for different years and months and blog entries and creates hundreds of errors. What's the best way to handle this for the SEOMoz report and the search engines. Should I rel=canonical the /blog page? I think this would probably affect the SEO of all the blog entries? Use robots.txt? Sitemaps? URL parameters in the search engines? Appreciate any assistance/recommendations Thanks in advance Ian
Technical SEO | | iragless0 -
What is the best practice to handle duplicate content?
I have several large sections that SEOMOZ is indicating has duplicate content, even though the content is not identical. For example: Leather Passport Section - Leather Passports - Black - Leather Passposts - Blue - Leather Passports - Tan - Etc. Each of the items has good content, but it is identical, since they are the same products. What is the best practice here: 1. Have only one product with a drop down (fear is that this is not best for the customer) 2. Make up content to have them sound different? 3. Put a do-no-follow on the passport section? 4. Use a rel canonical even though the sections are technically not identical? Thanks!
Technical SEO | | trophycentraltrophiesandawards0 -
Do dropdowns count as unique content?
My current site has some extensive unique database content by "widget" type. Currently we display this info into HTML 's, but we are considering utilizing this data in a dropdown field on each respective widget page. I want to ensure we don't have thin content...Does the content within the <option>tags on a dropdown count towards unique content?</option>
Technical SEO | | TheDude0 -
How critical is Duplicate content warnings?
Hi, So I have created my first campaign here and I have to say the tools, user interface and the on-page optimization, everything is useful and I am happy with SEOMOZ. However, the crawl report returned thousands of errors and most of them are duplicate content warnings. As we use Drupal as our CMS, the duplicate content is caused by Drupal's pagination problems. Let's say there is a page called "/top5list" , the crawler decided /top5list?page=1" to be duplicate of "/top5list". There is no real solution for pagination problems in Drupal (as far as I know). I don't have any warnings in Google's webmaster tools regarding this and my sitemap I submitted to Google doesn't include those problematic deep pages. (that are detected as duplicate content by SEOMOZ crawler) So my question is, should I be worried about the thousands of error messages in crawler diagnostics? any ideas appreciated
Technical SEO | | Gamer070 -
Where to put content on the page? - technical
The new algo update says any images at the top of the page negatively affect user experience if they are adverts? how does google know if its an advert or relevant banner? When trying to put text as far up as possible on the page, is it ok to make it appear higher in the code but appear further down using css? Or does Google not go from the code top to bottom when working this out, more how it renders? Any advice much appreciated.
Technical SEO | | pauledwards0