Duplicate Content
-
I was just using a program (copyscpape) to see if the content on a clients website has been copied. I was surprised that the content on the site was displaying 70% duplicated and it's showing the same content on a few sites with different % duplicated (ranging from 35%-80%) I have been informed that the content on the clients site is original and was written by the client.
My question is, does Google know or understand that the clients website's content was created as original and that the other sites have copied it word-for-word and placed it on their site?
Does he need to re-write the content to make it original? I just want to make sure before I told him to re-write all the content on the site? I'm well aware that duplicate content is bad, but i'm just curious if it's hurting the clients site because they originally created the content.
Thanks for your input.
-
It's absolutely ridiculous that our industry lays down and accepts stuff like this from Google. It is unacceptable for them to not be able to determine who originally created the content. They have the data to do it.
-
Thank you for the responses. They were both very helpful. I went ahead and took a few 10-word samples and searched it in Google with the quotation marks. The client showed up as the first result! He was pleased. Again, the responses were very helpful! Thanks.
I like that tips about using the quotations.
-
There is a way you can report the offending sites to Google if you can prove your content was created first.
https://support.google.com/webmasters/answer/93713?hl=en
If you are using a CMS, check the publish dates of the articles.
Like it has already been stated, do a search for some of the content by doing a search in Google for a section of content, surrounded by quotation marks. This tells Google to search for something exactly as it is entered, and return exact, not relevant results. If you see exact phrases that are duplicated word for word, you could contact either the site owner, or look up who hosts the site in Whois, and contact the hosting provider.
-
Take a few 10-word samples of the duplicated content and do a search for them in quotation marks. If your client's site shows up as the first result, you're in good shape. If it doesn't , strongly consider re-writing. Google does it's best to determine and rank the original source of content but, often, it gets it wrong and when that's the case, there isn't too much you can do about it other than building the authority/pagerank of the page from which the content was duplicated.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Does changing text content on a site affects seo?
HI, i have changed some h1 and h2 , changed and added paragraphs,fixed plagiarism,grammar and added some pics with alt text, I have just done it today, I am ranking on second page QUESTION-1 is it gonna affect my 2 months SEO efforts? QUESTION -2 Do I have to submit sitemap to google again? QUESTION-3 does changing content on the site frequently hurts SEO?
Algorithm Updates | | Sam09schulz0 -
How to unrank your content by following expert advice [rant]
Hi, As you can probably see from the title, a massive rant is coming up. I must admit I no longer understand SEO and I just wanted to see if you have any ideas what might be wrong. So, I read this blog post on MOZ https://moz.com/blog/influence-googles-ranking-factor - where the chap is improving ranking of content that is already ranking reasonably well. I've got two bits of news for you. The good news is - yes, you can change your articles' ranking in an afternoon. Bad news - your articles drop out of Top 100. I'll give you a bit more details hoping you can spot what's wrong. Disclaimer - I'm not calling out BS, I'm sure the blogger is a genuine person and he's probably has had success implementing this. The site is in a narrow but popular ecommerce niche where the Top 20 results are taken by various retailers who have simply copy/pasted product descriptions from the manufacturer's websites. The link profile strength is varied and I'm not making this up. The Top 20 sites range from DA:4 to DA:56. When I saw this I said to myself, it should be fairly easy to rank because surely the backlinks ranking factor weight is not as heavy in this niche as it is in other niches. My site is DA:18 which is much better than DA:4. So, even if I make my pages tiny tiny bit better than this DA:4 site, I should outrank it, right? Well, I managed to outrank it with really crap content. So, I got to rank two high-traffic keywords in #8 or #9 with very little effort. And I wish I stayed there because what followed just completely ruined my rankings. I won't repeat what was written in the blog. If you're interested, go and read it, but I used it as a blueprint and bingo, indeed Google changed my ranking in just a couple of hours. Wait, I lost more than 90 positions!!!! I'm now outside Top100. Now even irrelevant sites in Chinese and Russian are in front of me. They don't even sell the products. No, they're even in different niches altogether but they still outrank me. I now know exactly what Alice in Wonderland felt like. I want out please!!!!
Algorithm Updates | | GiantsCauseway0 -
Does cached duplicate content hurts seo by Google
If we have duplicate content or pages cached in Google which has been indexed months back, still it hurts the original pages? Old URLs with cache can be seen now in Google when we search for the same URLs.
Algorithm Updates | | vtmoz0 -
How much content is it safe to change?
I have read that it is unsafe to change more than 20% of your site’s content in any update. The rationale is that "Changing too much at once can flag your site within the Google algorithm as having something suspicious going on." Is this true, has anyone had any direct experiences of this or similar?
Algorithm Updates | | GrouchyKids0 -
Content on Wordpress blog inside the main website for SEO
Hi, We have our main website and our blog on blog.practo.com. Now what I see is that we wish to write in content to grow our seo keywords and links. Should we put the blog as www.practo.com/blog and then begin writing all the content or we should put the wordpress blog as www.practo.com/(wordpress blog here) and then begin writing the content. For best practices I suppose we should have content lined up as www.sitename.com/category/article name etc or www.sitename.com/article name etc - am I correct? Our main site consists of few html pages and then we have our software on a different sub domain. What are the best ways to publish content and get it crawled at a faster rate for growth? I would also wish to understand how to measure the number of growth in % to our content we are writing. Only via google analytics or some other tool? Say I wish to see the growth of 10 articles from month of may and compare it to the month of april or march 2012. So what tools could I use to see if we are progressing or not? Thanks
Algorithm Updates | | shanky10 -
What is considered duplicate content in an ecommerce website that offers the same product for retail and wholesale purchasing?
I have an ecommerce website that offers retail and wholesale products which are identical, of course with the exception of pricing. My concern is duplicate content. If the same product is offered under both the retail and wholesale category, and described identically, with the exception of price, metadata and a few words, is that considered duplicate content and would both pages be disregarded by the robots? Is it best to avoid the same description for that one product under the two separate categories? Thanks for all your help!
Algorithm Updates | | flaca0 -
Shouldn’t Google always rank a website for its own unique, exact +10 word content such as a whole sentence?
Hello fellow SEO's, I'm working with a new client who owns a property related website in the UK.
Algorithm Updates | | Qasim_IMG
Recently (May onwards) they have experienced significant drops in nearly all non domain/brand related rankings. From page 1 to +5 or worse. Please see the attached webmaster tools traffic graph.
The 13th of June seemed to have the biggest drop (UK Panda update???) When we copy and paste individual +20 word sentences from within top level content Google does bring up exact results, the content is indexed but the clients site nearly always appears at the bottom of SERP's. Even very new or small, 3-4 page domains that have clearly all copied all of their content are out ranking the original content on the clients site. As I'm sure know, this is very annoying for the client! And this even happens when Google’s cache date (that appears next to the results) for the clients content is clearly older then the other results! The only major activity was the client utilising Google optimiser which redirects traffic to various test pages. These tests finished in June. Details about the clients website: Domain has been around for 4+ years The website doesn't have a huge amount of content, around 40 pages. I would consider 50% original, 20% thin and 30% duplicate (working on fixing this) There haven’t been any signicant sitewide or page changes. Webmaster tools show nothing abnormal or any errors messages (some duplicate meta/title tags that are being fixed) All the pages of the site are indexed by Google Domain/page authority is above average for the niche (around 45 in for the domain in OSE) There are no ads of any kind on the site There are no special scripts or anything fancy that could cause problems I can't seem to figure it out, I know the site can be improved but such a severe drop where even very weak domains are out ranking suggests a penalty of some sort? Can anyone help me out here? hxuSn.jpg0 -
Is this the best way to get rid of low quality content?
Hi there, after getting hit by the Panda bear (30% loss in traffic) I've been researching ways to get rid of low quality content. From what I could find the best advise seemed to be a recommendation to use google analytics to find your worst performing pages (go to traffic sources - google organic - view by landing page). Any page that hasn't been viewed more than 100 times in 18 months should be a candidate for a deletion. Out of over 5000 pages and using this report we identified over 3000 low quality pages which I've begun exporting to excel for further examination. However, starting with the worst pages (according to analytics) I'm noticing some of our most popular pages are showing up here. For example: /countries/Panama is showing up as zero views but the correct version (with the end slash) countries/Panama/ is showing up as having over 600 views. I'm not sure how google even found the former version of the link but I'm even less sure how to proceed now (the webmaster was going to put a no-follow on any crap pages but this is now making him nervous about the whole process). Some advise on how to proceed from here would be fantastico and danke <colgroup><col width="493"></colgroup>
Algorithm Updates | | BrianYork-AIM0