What's the best way to manage content that is shared on two sites and keep both sites in search results?
-
I manage two sites that share some content. Currently we do not use a cross-domain canonical URL and allow both sites to be fully indexed. For business reasons, we want both sites to appear in results and need both to accumulate PR and other SEO/Social metrics. How can I manage the threat of duplicate content and still make sure business needs are met?
-
Does a duplicate content penalty impact specific pages or entire sites? If I wanted to test using the cross-domain canonical on a certain section of my site, would the impact be visible? Or would I need to put cross-domain canonicals on everything appearing on both sites in order to see the results?
-
Changing the articles or even page titles is not an option.
That's too bad. What Irving suggested has the potential for HUGE wins.
I'd find a way if that was my site.
-
Sure, that is a solution, but then rankings for the additional dupe sites went away because you basically suggested to Google "this URL on this site should not rank, because it is a copy of this article on this site, so give that site credit not me"
I believe that Jon has not been hit yet and wants both sites to rank, but is unable to change the content on either site to be unique. Any additional code you can insert in between the articles to create less similarity between both pages should help lessen the chance of getting hit but not a guarantee.
-
Irving, I had a client who had been hit with a manual penalty for Doorway Pages. They weren't Doorway Pages, they were just pages on various domains (that he owned) with a lot of duplicate content on them. We got him reinstated when we implemented cross-domain canonicals and filed a re-inclusion request. Sounds similar to this case?
Just wondering if anyone had heard of sites being hit like that for dupe content?
-
LOL true.
With all due respect, 301, noindex or cross-canonicalizing is as much of a solution as saying delete your second site. My suggestion of breaking up the content or appending additional content will possibly help you avoid a dupe content filter being triggered.
Duplicate content is not a penalty, it's a filter so the worst that happens is the main site that was bringing you the majority of traffic gets filtered and loses rankings to the secondary site.
I think a good question to ask at this point would be for you to clarify your first sentence: "I manage two sites that share some content" can you define what "some" means? are they main conversion pages or secondary blog posts, and what percentage of the site is dupe content?
BTW, hope you're not interlinking your two sites keep them as separate as possible.
-
Try this post for more info:
http://googlewebmastercentral.blogspot.com/2009/12/handling-legitimate-cross-domain.html
-
Sounds like you don't need to manage the threat of duplicate content; you are producing the duplicate content yourself. You are instead wanting to minimize the effect duplicate content has from one site to the next. The only way I know of to get eliminate the risk of duplicate content penalties is to noindex, 301 redirect, or provide canonical URLs.
Since you want both sites to continue being indexed, you can either keep doing what you're doing (and hope you don't get hit) or use canonical URLs and pick which site is best for each page.
Hope this helps.
-
If I used the cross-domain canonical, would that mean that one site would stop appearing in search results?
-
You can append additional content to the bottom of the page on the more important site, or break up the article by adding content and or ads between the paragraphs (which will probably result in article fragmentation) but if you're not a news source it's not a big deal.
-
I'm no technical expert but it sounds like you're playing with fire. I've seen more than one site penalised for exactly this. If it looks like you're trying to rank the same piece of content twice, at least one of the URLs is at risk of filtering or a penalty. Isn't this exactly what the cross-domain canonical was created for?
-
Changing the articles or even page titles is not an option.
-
Paraphrase the articles on the highest traffic pages to your secondary site and/or tweak the keyword targets
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What's the best redirect to use for a newer version of a blog post?
For example: suppose you have a post "The Best Games to Play for YouTube Gamers in 2016" and you want to make this a yearly series. Should you 301 the 2016 version to the new 2017 one? Should you use the canonical attribute? If 2016 isn't in the URL, should you make the 2017 one the new URL?
Intermediate & Advanced SEO | | Edward_Sturm0 -
301 redirect to search results page?
Hi - we just launched our redesigned website. On the previous site, we had multiple .html pages that contained links to supporting pdf documentation. On this new site, we no longer have those .html landing pages containing the links. The question came up, should we do a search on our site to gather a single link that contains all pdf links from the previous site, and set up a redirect? It's my understanding that you wouldn't want google to index a search results page on your website. Example: old site had the link http://www.oldsite.com/technical-documents.html new site, to see those same links would be like: http://www.newsite.com/resources/search?View+Results=&f[]=categories%3A196
Intermediate & Advanced SEO | | Jenny10 -
Will disallowing URL's in the robots.txt file stop those URL's being indexed by Google
I found a lot of duplicate title tags showing in Google Webmaster Tools. When I visited the URL's that these duplicates belonged to, I found that they were just images from a gallery that we didn't particularly want Google to index. There is no benefit to the end user in these image pages being indexed in Google. Our developer has told us that these urls are created by a module and are not "real" pages in the CMS. They would like to add the following to our robots.txt file Disallow: /catalog/product/gallery/ QUESTION: If the these pages are already indexed by Google, will this adjustment to the robots.txt file help to remove the pages from the index? We don't want these pages to be found.
Intermediate & Advanced SEO | | andyheath0 -
Sitemaps during a migration - which is the best way of dealing with them?
Many SEOs I know simply upload the new sitemap once the new site is launched - some keep the old site's URLs on the new sitemap (for a while) to facilitate the migration - others upload both the old and the new website together, to support the migration. Which is the best way to proceed? Thanks, Luke
Intermediate & Advanced SEO | | McTaggart0 -
How do I prevent 404's from hurting my site?
I manage a real estate broker's site on which the individual MLS listing pages continually create 404 pages as properties are sold. So, on a site with 2200 pages indexed, roughly half are 404s at any given time. What can I do to mitigate any potential harm from this?
Intermediate & Advanced SEO | | kimmiedawn0 -
Site Structure: How do I deal with a great user experience that's not the best for Google's spiders?
We have ~3,000 photos that have all been tagged. We have a wonderful AJAXy interface for users where they can toggle all of these tags to find the exact set of photos they're looking for very quickly. We've also optimized a site structure for Google's benefit that gives each category a page. Each category page links to applicable album pages. Each album page links to individual photo pages. All pages have a good chunk of unique text. Now, for Google, the domain.com/photos index page should be a directory of sorts that links to each category page. Alternatively, the user would probably prefer the AJAXy interface. What is the best way to execute this?
Intermediate & Advanced SEO | | tatermarketing0 -
Our Site's Content on a Third Party Site--Best Practices?
One of our clients wants to use about 200 of our articles on their site, and they're hoping to get some SEO benefit from using this content. I know standard best practices is to canonicalize their pages to our pages, but then they wouldn't get any benefit--since a canonical tag will effectively de-index the content from their site. Our thoughts so far: add a paragraph of original content to our content link to our site as the original source (to help mitigate the risk of our site getting hit by any penalties) What are your thoughts on this? Do you think adding a paragraph of original content will matter much? Do you think our site will be free of penalty since we were the first place to publish the content and there will be a link back to our site? They are really pushing for not using a canonical--so this isn't an option. What would you do?
Intermediate & Advanced SEO | | nicole.healthline1 -
Best way to de-index content from Google and not Bing?
We have a large quantity of URLs that we would like to de-index from Google (we are affected b Panda), but not Bing. What is the best way to go about doing this?
Intermediate & Advanced SEO | | nicole.healthline0