Pages mirrored on unknown websites (not just content, all the HTML)... blackhat I've never seen before.
-
Someone more expert than me could help... I am not a pro, just doing research on a website... Google Search Console shows many backlinks in pages under unknown domains... this pages are mirroring the pages of the linked website... clicking on a link on the mirror page leads to a spam page with link spam... The homepage of these unknown domain appear just fine... looks like that the domain is partially hijacked... WTF?!
Have you ever seen something likes this?
Can it be an outcome of a previous blackhat activity?
-
The links actually are just in some forms
-
Thanks everyone, I'll do more research and update all
-
One more thing that could be a long shot but is worth keeping an eye on: If someone had zero ethics and wanted to steal your links, one thing they could do is reach out to webmasters that link to you and pretend to be you, and claim that the mirror site is the new version of the website and ask them to change their links. Once complete, they'd redirect the mirror site to their own commercial site.
So, keep an eye on your own link profile and theirs, just to make sure none of those links are changing.
-
We see scraped content all the time. Occasionally we see scraped HTML in full form but slightly edited.
In general I'd do the following:
- Disavow those domains in GSC. Also consider disavowing any domains pointing links at the mirror sites. It could be possible that they are building up junk links to a mirror site which canonicals your site. No idea why anyone would do that but worth covering your bases.
- Submit DMCA notices to Google, their webhost, etc.
- Take a look through server logs for anything screwy. If you see tons of weird traffic from countries that aren't relevant to your business, you might decide to block them with your .htaccess file.
- Certainly make sure nothing has been changed on your own site (canonicals, common Wordpress hacks, etc.), though nothing is suggesting to me that that has happened.
Other than that I would just keep an eye on it and keep an eye on Google Search Console.
-
Hi,
From what I read it looks similar to these cases - https://moz.com/community/q/getting-different-search-queries-in-google-webmaster & http://moz.com/community/q/chinese-site-ranking-for-our-brand-name-possible-hack
In these cases - sites where hacked using a vulnerability in a Wordpress slider - they copied sites they wanted to target on to some of the hacked domains and used other hacked domains to point links to them. At that point they were trying to redirect traffic from genuine e-commerce sites to bogus ones.
You could try to contact the site owners & tell them they have been hacked. At the same time file a complaint @Google
rgds,
Dirk
-
Some basic scraping maybe common, but they have gone out of there way to backlink to you...
Have you checked there metrics on open site explore...? Checked whois?
-
As far as I know website scraping is not uncommon, apart from the backlinks
-
I am honest I am still a little puzzled. But if what you are describing is what is happening. Something very strange is going on. I have not heard of it before.
A totally different domain, that you did not build is linking to your site with the exact copy of a page that it is linked to. If that is the case, your site is being specifically targeted. Someone has to have gone out of there way to copy your site and then re-build it on a different domain.
This is what I would do:-
I would contacting google asap.
Can you find who owns the site on whois?
Check the wayback machine to find out what the site was previously
User open site explorer to track the nature of the links on the other domain. .
And wow... that is unbelievable if it is correct...
-
Yes, sorry but I thought that was clear from the beginning.
If I click to the console links I am directed to the mirrored page in these domains. That's exactly how I saw them, otherwise I coudn't discover them
-
That console part provides clickable links. So that goes back to my original question. Can you click through to the domain and see a mimic copy of your site? Or is it simply showing you a link that clicks to nothing?
-
In the console there is a section called (translating into English) "Links pointing to your site"
Traffic has not changed, in fact these do not appear in GA referrals.
I can see all these domains, either the mirrorer paged, or the actual genuine pages pertaining to them originally. You cannot reach the mirrored pages via the menu in hp
-
Where in google search console are they reported?
Also to be clear, nothing has happened to your site. ie traffic has not changed. However you have found "strange" links in search console? Which do not click through to anything?
Or as I am still not clear - have you found another domain, that mimics your site that you can actually see?
-
Exactly, I am confused too.
Indeed, Goole Search Console reports do list them, but I can't see them in the html
Instance, I find this
<form <span="" class="html-tag" data-mce-mark="1">id="search_mini_form" action="http://mydomain.ext/it/catalogsearch/result/" method="get">
and also images hosted in the scraped websites's domain
Also, these pages apparently are under these domains that in turn appear to be real websites equally unaware of the scam
</form>
-
I am a little confused.
Are you saying that an exact copy of a unique page you have created on your website is on another domain name (which you know nothing about) and then that site is linked to your site, ie via a backlink?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Third part http links on the page source: Social engineering content warning from Google
Hi, We have received "Social engineering content" warning from Google and one of our important page and it's internal pages have been flagged as "Deceptive site ahead". We wonder what's the reason behind this as Google didn't point exactly to the specific part of the page which made us look so to the Google. We don't employ any such content on the page and the content is same for many months. As our site is WP hosted, we used a WordPress plugin for this page's layout which injected 2 http (non-https) links in our page code. We suspect if this is the reason behind this? Any ideas? Thanks
White Hat / Black Hat SEO | | vtmoz1 -
Do the links from top websites' forums boost in-terms of backlinks?
If we get any backlinks from discussions/forums of top websites like wordpress and joomla forums; do they count as valid and authority improving backlinks? I mean about the dofollow links.
White Hat / Black Hat SEO | | vtmoz1 -
Canonical tags being direct to "page=all" pages for an Ecommerce website
I find it alarming that my client has canonical tags pointing to "page=all" product gallery pages. Some of these product gallery pages have over 100 products and I think this could effect load time, especially for mobile. I would like to get some insight from the community on this, thanks!
White Hat / Black Hat SEO | | JMSCC0 -
Looking for a Way to Standardize Content for Thousands of Pages w/o Getting Duplicate Content Penalties
Hi All, I'll premise this by saying that we like to engage in as much white hat SEO as possible. I'm certainly not asking for any shady advice, but we have a lot of local pages to optimize :). So, we are an IT and management training course provider. We have 34 locations across the US and each of our 34 locations offers the same courses. Each of our locations has its own page on our website. However, in order to really hone the local SEO game by course topic area and city, we are creating dynamic custom pages that list our course offerings/dates for each individual topic and city. Right now, our pages are dynamic and being crawled and ranking well within Google. We conducted a very small scale test on this in our Washington Dc and New York areas with our SharePoint course offerings and it was a great success. We are ranking well on "sharepoint training in new york/dc" etc for two custom pages. So, with 34 locations across the states and 21 course topic areas, that's well over 700 pages of content to maintain - A LOT more than just the two we tested. Our engineers have offered to create a standard title tag, meta description, h1, h2, etc, but with some varying components. This is from our engineer specifically: "Regarding pages with the specific topic areas, do you have a specific format for the Meta Description and the Custom Paragraph? Since these are dynamic pages, it would work better and be a lot easier to maintain if we could standardize a format that all the pages would use for the Meta and Paragraph. For example, if we made the Paragraph: “Our [Topic Area] training is easy to find in the [City, State] area.” As a note, other content such as directions and course dates will always vary from city to city so content won't be the same everywhere, just slightly the same. It works better this way because HTFU is actually a single page, and we are just passing the venue code to the page to dynamically build the page based on that venue code. So they aren’t technically individual pages, although they seem like that on the web. If we don’t standardize the text, then someone will have to maintain custom text for all active venue codes for all cities for all topics. So you could be talking about over a thousand records to maintain depending on what you want customized. Another option is to have several standardized paragraphs, such as: “Our [Topic Area] training is easy to find in the [City, State] area. Followed by other content specific to the location
White Hat / Black Hat SEO | | CSawatzky
“Find your [Topic Area] training course in [City, State] with ease.” Followed by other content specific to the location Then we could randomize what is displayed. The key is to have a standardized format so additional work doesn’t have to be done to maintain custom formats/text for individual pages. So, mozzers, my question to you all is, can we standardize with slight variations specific to that location and topic area w/o getting getting dinged for spam or duplicate content. Often times I ask myself "if Matt Cutts was standing here, would he approve?" For this, I am leaning towards "yes," but I always need a gut check. Sorry for the long message. Hopefully someone can help. Thank you! Pedram1 -
If our site hasn't been hit with the Phantom Update, are we clear?
Our SEO provider created a bunch of "unique url" websites that have direct match domain names. The content is pretty much the same for over 130 websites (city name is different) that link directly to our main site. For me this was a huge red flag, but when I questioned them and they said it was fine. We haven't seen a drop in traffic, but concerned that Google just hasn't gotten to us. DA for each of these sites are 1 after several months. Should we be worried? I think yes, but I am an SEO newbie.
White Hat / Black Hat SEO | | Buddys0 -
Website mallware attacks
I keep getting attacks to my website every time that are being blocked by OSE firewall Is there any way to stop this? I am affraid because they actually manage enter my website on the past, and i dont know if they can enter on the future or if having all the pluggins and wordpress updated. I am safe enough, and i am not sure if there is any type of virus on my computer Macbook as those attacked pages were recently updated from my computer. Is there any malware scan for Mac Thanl you == Attack Details == TYPE: Found Basic DoS Attacks DETECTED ATTACK VALUE: dDos Attack ACTION: Blocked LOGTIME: 2013-02-25 11:48:18 FROM IP: http://whois.domaintools.com/75.126.24.81 URI: [http://www.propdental.es/](http://www.propdental.es/) METHOD: HEAD USERAGENT: N/A REFERRER: N/A == Attack Details == TYPE: Found Basic DoS Attacks DETECTED ATTACK VALUE: dDos Attack ACTION: Blocked LOGTIME: 2013-02-25 10:13:17 FROM IP: http://whois.domaintools.com/107.21.150.82 URI: [http://www.propdental.es/blanqueamiento-dental/](http://www.propdental.es/blanqueamiento-dental/) METHOD: HEAD USERAGENT: N/A REFERRER: N/A ``` == Attack Details == TYPE: Found Malicious User Agent DETECTED ATTACK VALUE: curl/7.15.5 (x86_64-redhat-linux-gnu) libcurl/7.15.5 OpenSSL/0.9.8b zlib/1.2.3 libidn/0.6.5 ACTION: Blocked LOGTIME: 2013-02-25 03:13:52 FROM IP: http://whois.domaintools.com/119.245.226.74 URI: [http://www.propdental.es/sonrisas/los-martinez/](http://www.propdental.es/sonrisas/los-martinez/) METHOD: HEAD USERAGENT: curl/7.15.5 (x86_64-redhat-linux-gnu) libcurl/7.15.5 OpenSSL/0.9.8b zlib/1.2.3 libidn/0.6.5 REFERRER: N/A ``` ```
White Hat / Black Hat SEO | | maestrosonrisas0 -
Dealing with internal pages with bad backlinks - is this approach OK?
Hi all, I've just been going through every page of my company website, and found a couple of internal pages with nasty backlinks/profiles. There are a significant number of article marketing and rubbish directory pages pointing to these internal pages. These internal pages have low PR, yet are performing well in terms of SERPs. I was planning to: (1) change URLs - removing current (soon to be former) URLs from Google via Webmaster Tools. Then (2) remove website's 404 for a while so nasty links aren't coming anywhere near the website (hopefully nasty links will fail to find website and broken links will result in link removal - that's my thinking anyway). PS. I am not planning to implement any kind of redirect from the old URLs. Does this sound like a sensible approach, or may there be problems with it? Thanks in advance, Luke
White Hat / Black Hat SEO | | McTaggart0 -
Competitors Developing Spammy Link For My Website
Well Guys there are lot of discussions in almost all the communities, blogs, forums about Post Penguin impact. Google says that if find that you're involved in any link building activities, we may penalize you. People out there have already started their developed links. But what if our competitors would have developed those links. Initially it was okay to develop one way links, I even developed lot of quality, but deliberately, links. around 95% links are placed manually, if return to some favor or money but all links looks natural. Most of the links I developed through content only, like articles, blog comments, PR submission, etc now really skeptical about the quality (after hearing lot of talks and reading n number of posts). Now, can I also submit my competitor's websites in 1000 topic directory (obviously not in any spammy directory), would it effect that website adversely? What if I spun an existing content and submit it into 500 article directories and give backlink to competitor site from using only one anchor text (which is obviously the main keywords - highest sales generating keyword) I look forward to some experts comments.
White Hat / Black Hat SEO | | Khem_Raj70