Duplicate content issue, across site domains (blogging)
-
Hi all,
I've just come to learn that a client has been cross-posting their blog posts to other blogs (on higher quality domains, in some cases). For example - this is the same post on 3 different blogs.
http://thebioethicsprogram.wordpress.com/2014/06/30/how-an-irb-could-have-legitimately-approved-the-facebook-experiment-and-why-that-may-be-a-good-thing/
http://blogs.law.harvard.edu/billofhealth/2014/06/29/how-an-irb-could-have-legitimately-approved-the-facebook-experiment-and-why-that-may-be-a-good-thing/
http://www.thefacultylounge.org/2014/06/how-an-irb-could-have-legitimately-approved-the-facebook-experimentand-why-that-may-be-a-good-thing.html
And, sometimes a 4th time, on an NPR website.I'm assuming this is doing no one any favors and Harvard or NPR is going to earn the rank most every time. I'm going to encourage them to publish only fresh content on their real blog, would you agree? Can this actually harm the ranking of their blog and website - should we delete the old entries when migrating the blog? They are going to move their Wordpress Blog to hosting on their real domain soon:
http://www.bioethics.uniongraduatecollege.edu/news/The current set up is not adding any value to their domain. Thank you for any advice!
Ketan
-
Ketan,
I'm going to encourage them to publish only fresh content on their real blog, would you agree?
If you look at pretty much any of the blog posts on these forums you will see that more and more everything comes back to content. Original content. Original Content. Did I mention original content? EGOL shared a link with me and I'm sharing it again on another post but it's pertinent.
http://www.thesempost.com/google-rewrites-quality-rating-guide-seos-need-know/
Content used to be king and he wants his thrown back!
Can this actually harm the ranking of their blog and website - should we delete the old entries when migrating the blog?
Duplicate content is something that I am working a lot with right now. There is a difference between duplicate content and plagiarism. What needs to be determined is how the are using this content. Matt Cutts said that roughly 25% of the internet is duplicate content, and a lot of duplicate content is ok, for example if you were writing an analysis, or writing commentary on one of those blog posts then of course there is going to be some duplicate content. That sort of this would be ok.
If they are simply taking the article, and posting it, yet still giving credit, then no its not really doing anything except potentially giving your blog traffic. Take social bookmarking websites for example. Let's look at digg.com. A long time ago digg used to take the entire article and post it to their page and you could actually view everything within their website and everything was all gravy. Now, that doesn't work so well. Now, digg usually writes a little blurb, and provides a link.
Google will look at duplicate content, determine which ever is the best representation of the content, usually who wrote it first, who has the strongest domain etc, and gives the credit to them. The other thing too is that these websites, Havard, NPR, etc have in their favor is that they are prob indexed often enough to guarantee that they are going to get credit first.
As to the question about deleting them. I don't think you necessarily need to delete them, depending on how many articles exist, how much traffic they generate etc. There is a lot to look at. If nobody is looking at them, then sure, you can always do a 301 to one of your new blog posts later down the line. Or maybe your first posts are rewrites or analysis of the articles. Regardless I would provide a link on all of the pages letting people know where you got the information, that way nobody can say you were trying to steal the information.
My thought on this whole thing. If it makes you uncomfortable, it's gonna make Google feel uncomfortable.
Hope that helps! Good luck!
Matt Cutts on Duplicate content: https://www.youtube.com/watch?v=mQZY7EmjbMA&feature=kp
Matt Cutts on Original content: https://www.youtube.com/watch?v=4LsB19wTt0Q
-
Do they use canonical URLs on their website? Do the cross-posted blog posts have canonicals that point back to the originating site? If so, the duplicate content should not harm their website and their website should rank above the bigger players for their content (though not necessarily--Google works in mysterious ways).
In general, it is a good idea to keep your best, fresh content on your own website but sometimes it can be good to get the added exposure of a high-traffic website, assuming that it is clear where the content originated.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
On Site Question: Duplicate H2...
Hi All A few on-site audit tools pull information on duplicate H2 tags on pages. This implies it's a bad thing and should be fixed - is that the case? On one of my sites the tag-line is in H2 in the header, so appears on every page... Just wondering if this is something worth fixing. Thanks
On-Page Optimization | | GTAMP0 -
Robot.txt file issue on wordpress site.
I m facing the issue with robot.txt file on my blog. Two weeks ago i done some development work on my blog. I just added few pages in robot file. Now my complete site seems to be blocked. I have checked and update the file and still having issue. The search result shows that "A description for this result is not available because of this site's robots.txt – learn more." Any suggestion to over come with this issue
On-Page Optimization | | Mustansar0 -
Duplicate Content - But it isn't!
Hi All, I have a site that releases alerts for particular problem/events/happenings. Due to legal stuff we keep the majority of the content the same on each of these event pages. The URLs are all different but it keeps coming back as duplicate content. The canonical tag is not right (i dont think for this) egs http://www.holidaytravelwatch.com/alerts/call-to-arms/egypt/coral-sea-waterworld-resort-sharm-el-sheikh-egypt-holiday-complaints-july-2014 http://www.holidaytravelwatch.com/alerts/call-to-arms/egypt/hotel-concorde-el-salam-sharm-el-sheikh-egypt-holiday-complaints-may-2014
On-Page Optimization | | Astute-Media0 -
Duplicate Page Content
Hey Moz Community, Newbie here. On my second week of Moz and I love it but have a couple questions regarding crawl errors. I have two questions: 1. I have a few pages with duplicate content but it say 0 duplicate URL's. How do I know what is duplicated in this instance? 2. I'm not sure if anyone here is familiar with an IDX for a real estate website. But I have this setup on my site and it seems as though all the links it generates for different homes for sale show up as duplicate pages. For instance, http://www.handyrealtysa.com/idx/mls...tonio_tx_78258 is listed as having duplicate page content compared with 7 duplicate URLS: http://www.handyrealtysa.com/idx/mls...tonio_tx_78247
On-Page Optimization | | HandyRealtySA
http://www.handyrealtysa.com/idx/mls...tonio_tx_78253
http://www.handyrealtysa.com/idx/mls...tonio_tx_78245
http://www.handyrealtysa.com/idx/mls...tonio_tx_78261
http://www.handyrealtysa.com/idx/mls...tonio_tx_78258
http://www.handyrealtysa.com/idx/mls...tonio_tx_78260
http://www.handyrealtysa.com/idx/mls...tonio_tx_78260 I've attached a screenshot that shows 2 of the pages that state duplicate page content but have 0 duplicate URLs. Also you can see somewhat about the idx duplicate pages. rel="canonical" is functioning on these pages, or so it seems when I view the source code from the page. Any help is greatly appreciated. skitch.png0 -
Redirect closing established site to another established site.
We have 2 ecommerce websites with duplicate product/category content that have been around for years. "Site A" since 2004 and "Site B" since 2011. "Site B" doesn't generate many sells and we want to close the site in fear that it may look spammy to google and since we don't want to upkeep the site. Couple questions:
On-Page Optimization | | Tuurbo
1. Is it possible that "Site B" has damaged "Site A" with google since they are so similar?
2. Can i 301 redirect all of "Site B" to "Site A" without hurting "Site A"?
3. If question 2 is ok, should i use the "Change of Address" tool in google webmaster tools to point "Site B" to "Site A"? Thanks!0 -
Checking for content originality in a site
two part question on original content How would you go about checking if a site holds original content accept the long search quary within Google? ans also if I find many sites carrying my content and I am the original source should I replace the content? thanks
On-Page Optimization | | ciznerguy0 -
The crawl diagnosis indicated that my domain www.mydomain.com is duplicate with www.mydomain.com/index.php. How can I correct this issue?
How can I fix this issue when crawl diagnosis indicated that my www.mydomain.com is duplicate with www.mydomain.com/index.php? That suppose to be the same page and not duplicate, right?
On-Page Optimization | | jsevilla0 -
What is the best way to manage industry required duplicate Important Safety Information (ISI) content on every page of a site?
Hello SEOmozzer! I have recently joined a large pharmaceutical marketing company as our head SEO guru, and I've encountered a duplicate content related issue here that I'd like some help on. Because there is so much red tape in the pharmaceutical industry, there are A LOT of limitations on website content, medication and drug claims, etc. Because of this, it is required to have Important Safety Information (ISI) clearly stated on every page of the client's website (including the homepage). The information is generally pretty lengthy, and in some cases is longer than the non-ISI content on each page. Here is an example: http://www.xifaxan.com/ All content under the ISI header is required on each page. My questions are: How will this duplicated content on each page affect our on-page optimization scores in the eyes of search engines? Is Google seeing this simply as duplicated content on every page, or are they "smart" enough to understand that because it is a drug website, this is industry standard (and required)? Aside from creating more meaty, non-ISI content for the site, are there any other suggestions you have for handling this potentially harmful SEO situation? And in case you were going to suggest it, we cannot simply have an image of the content, as it may not be visible by all internet users. We've already looked into that 😉 Thanks in advance! Dylan
On-Page Optimization | | MedThinkCommunications0