Site scraped over 400,000 urls
-
Our business is heavily dependent on SEO traffic from long tail search. We have over 400,000 pieces of content, all of which we found scraped and published by another site based out of Hong Kong (we're in the US).
Google has a process for DMCA takedown, but doing so would be beyond tedious for such a large set of urls. The scraped content is outranking us in many searches and we've noticed a drastic decrease in organic traffic, likely from a duplicate content penalty.
Has anyone dealt with an issue like this? I can't seem to find much help online.
-
Hi Kibin
Firstly it's unlikely that their scraped content will affect your rankings - Google generally knows who originated it. However:
Do you have the hreflang tag on your website? specifying your language and location? If theirs has this as well then technically you are targetting a different country, so there should be no duplicate content if you added it.
https://support.google.com/webmasters/answer/189077?hl=en
I would tell Google about the URL and add a sample 10 URLs first: https://www.google.com/webmasters/tools/dmca-dashboard. Telling them is an absolute must even if it's only a few URLs.
Also email the hosting company informing them that they are hosting copied content and that the penalties are severe.
Finally, write to the company themselves and tell them/warn them that you are going legal and send them a cease and desist legal letter. I am sure you can knock one up for a few dollars from a friendly solicitor.
Watch this: https://www.youtube.com/watch?v=gGc_jc3Oznk It's a bit long but worth it.
Do all of these things.
Regards
Nigel
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
URLs too long, but run an eCommerce site
Hi, When I started out I was pretty green to SEO, and didn't consider the usability/SEO impact of URL structure. Flash forward, I'm 5 years deep into using the following: mysite.com/downloads/category/premium-downloads/sub-category/ ("category" is quite literally one rung on the link - thanks, WordPress - however "sub-category" is a placeholder) I run a digital downloads store, and I now have 100s ofinternal links beholden to this hideous category linking structure. Not to mention external links at Google Ads, etc. I would LOVE to change this, but if I were to do so, what should I consider? For instance, is there a checklist for making a change like this? I was thinking of changing it to something like the following: mysite.com/shop/c/premium/sub-category/ And also, how much damage, if any, would this be doing to my SEO? Thanks in advance,
Technical SEO | | LouCommaTheCreator
Lou1 -
URL is invalid: Why?
Hello everyone, I am currently listing my company on business directories. For some websites however when I add my website URL, it comes up as URL is invalid. What could be the reason for this? I have tried different variations like www., http:// and https://. Kind Regards,
Technical SEO | | SMCCoachHire
Aqib0 -
How to create site map for large site (ecommerce type) that has 1000's if not 100,000 of pages.
I know this is kind of a newbie question but I am having an amazing amount of trouble creating a sitemap for our site Bestride.com. We just did a complete redesign (look and feel, functionality, the works) and now I am trying to create a site map. Most of the generators I have used "break" after reaching some number of pages. I am at a loss as to how to create the sitemap. Any help would be greatly appreciated! Thanks
Technical SEO | | BestRide0 -
Site Indexed but not Cached?
I launched a new website ~2 weeks ago that seems to be indexed but not cached. According to Google Webmaster most of the pages are indexed and I see them appear when I search site:www.xxx.com. However, when I type into the URL - cache:www.xxx.com I get a 404 error page from Google.
Technical SEO | | theLotter
I've checked more established websites and they are cached so I know I am checking correctly here... Why would my site be indexed but not in the cache?0 -
Poor Site Performance
Hello, A couple of months ago, this site was dropped from google due to a noindex, nofollow tag thewealthymind(dot)com It's back up, but performing poorly. Take for example the term "The 4 step belief change" in the home page title tag. This site is the #1 authority on that and yet it ranks 3rd below weaker pages. There's 180 404 errors in GWT, many from past versions of pages of the site but also including thewealthymind(dot)com/index.html and thewealthymind(dot)com/index.htm even though there is a rel=cononical tag on the home page. What's the process of getting this site back to health?
Technical SEO | | BobGW0 -
Multiple URLs
I'm trying to check the URLs of this site- http://www.ofo.com.au, and I see that their old site has 301 re-directed to it...but the site http://ofo.com.au and http://outdoorfurnitureoutlet.com.au are both still up and I can't see any 301 redirects from them. Is it a problem even if when I do a site: search for them I get no results?
Technical SEO | | UnaRealidad0 -
Overly Dynamic URLs
I have a site that I use to time fitness events and I like to post the results using query strings. I create a link to each event's results/gallery/etc. I don't need these pages crawled and I don't want them to hurt my seo. Can I put a "do not crawl" meta on them or will that hurt my overall positioning? What are my other options?
Technical SEO | | bobbabuoy0 -
Redirecting a old aged site to a new exact match site?
Hi All, I have a question. I have 2 sites with me in the same sector and want some help. site 1 is a old site started back in 2003 and has some amount of links to it and has a pr 3 with some good links to it but doesn't rank much for any keywords for the timing. site 2 is a aged domain but newly developed with unique content and has a good amount of exact match with a .com version. so will there be any benefit by redirecting site 1 to site 2 to get the seo benefits and a start for link bulding? or is it best to develop and work on each site? the sector is health insurance. Thanks
Technical SEO | | macky71