A doorway-page vendor has made my SEO life a nightmare! Advice anyone!?
-
Hey Everyone,
So I am the SEO at a mid-sized nationwide retailer and have been working there for almost a year and half. This retailer is an SEO nightmare. Imagine the worst possible SEO nightmare, and that is my unfortunate yet challenging everyday reality.
In light of the new algorithm update that seems to be on the horizon from Google to further crack down on the usage of doorway pages, I am coming to the Moz community for some desperately needed help.
Before I was employed here, the eCommerce director and SEM Manager connected with a vendor that told them basically that they can do a PPC version of SEO for long-tail keywords. This vendor sold them on the idea that they will never compete with our own organic content and can bring in incremental traffic and revenue due to all of this wonderful technology they have that is essentially just a scraper.
So for the past three years, this vendor has been creating thousands of doorway pages that are hosted on their own server but our masked as our own pages. They do have a massive index / directory in HTML attached to our website and even upload their own XML site maps to our Google Web Master Tools. So even though they “own” the pages, they masquerade as our own organic pages.
So what we have today is thousands upon thousands of product and category pages that are essentially built dynamically and regurgitated through their scraper / platform, whatever.
ALL of these pages are incredibly thin in content and it’s beyond me how Panda has not exterminated them.
ALL of these pages are built entirely for search engines, to the point that you would feel like the year was 1998.
All of these pages are incredibly over- optimized with spam that really is equivalent to just stuffing in a ton of meta keywords. (like I said – 1998)
Almost ALL of these scraped doorway pages cause an incredible amount of duplicate content issues even though the “account rep” swears up and down to the SEM Manager (who oversees all paid programs) that they do not.
Many of the pages use other shady tactics such as meta refresh style bait and switching.
For example:
The page title in the SERP shows as: Personalized Watch Boxes
When you click the SERP and land on the doorway page the title changes to:
Personalized Wrist Watches. Not one actual watch box is listed.
They are ALL simply the most god awful pages in terms of UX that you will ever come across BUT because of the sheer volume of this pages spammed deep within the site, they create revenue just playing the odds game.
Executives LOVE revenue.
Also, one of this vendor’s tactics when our budget spend is reduced for this program is to randomly pull a certain amount of their pages and return numerous 404 server errors until spend bumps back up. This causes a massive nightmare for me.
I can go on and on but I think you get where I am going.
I have spent a year and half campaigning to get rid of this black-hat vendor and I am finally right on the brink of making it happen. The only problem is, it will be almost impossible to not drop in revenue for quite some time when these pages are pulled. Even though I have helped create several organic pages and product categories that will pick-up the slack when these are pulled, it will still be awhile before the dust settles and stabilizes.
I am going to stop here because I can write a novel and the millions of issues I have with this vendor and what they have done. I know this was a very long and open-ended essay of this problem I have presented to you guys in the Moz community and I apologize and would love to clarify anything I can.
My actual questions would be:
Has anyone gone through a similar situation as this or have experience dealing with a vendor that employs this type of black-hat tactic?
Is there any advice at all that you can offer me or experiences that you can share that can help be as armed as I can when I eventually convince the higher-ups they need to pull the plug?
How can I limit the bleeding and can I even remotely rely on Google LSI to serve my organic pages for the related terms of the pages that are now gone?
Thank you guys so much in advance,
-Ben
-
glad to help
-
You are a genius.
-
Glad i could be of some help,
If I were you I'd definitely grab copies of the pages if they're still live, you could do this from home even using some free tools like
http://phpcrawl.cuab.de/about.html
add a bit of Curl or WGET and you've got the pages plus the links and meta. Then if they do disappear suddenly and the business is stuck, you can hand this to your web people at oracle and they'll probably try and hire you, having said that, I'd imagine they've probably got a decent contingency plan because they're oracle, but you never know. Could save the day.
-
Thanks Jamie!
Yea we actually partner with Oracle for our web design, engineering , implementation and so on. So when it comes to server-side issues, we would have to go through them and there is always red tape involved.
Really I cannot understand how this vendor that does this is even in business and it is beyond me how they even get away with it. The wordpress 404 plug-in is a great idea though and that will definitely help me in the future with freelancing while I am here full-time.
-
We do self-canocialize and that is a very good question. What they will do is just keep spitting out dynamically generated URLS. They have absolutely no restrictions on page quality, content, they literally have no rules. This gives them immense flexibility.
And for the contract portion: One the contract ends, all of these pages will in-fact disappear and that is why they house them on their own servers. So that is what we want in the end.
It is dealing with the massive amount of 404s that will be an issue for awhile.
-
Thanks again!
Yes, that is the conundrum I am in here when it comes to "who actually owns the pages" and honestly, this vendor covered their bases. They actually house all of the pages on their own servers and basically scrape out site, then shoot them out through our CDN via a proxy or something like that. So they made sure we are at their mercy, they can pull them anytime they want.
So technically, If were were to redirect all of their pages and acquired links, it would actually not be too hard because each page is so unbelievably identical to our own organic pages. The problem is, we would have to access their server I believe and that will not happen.
It will also be one hell of a mess with 301s if we were to do that and I know someone I am planning with on our site team fears the length of the 301 chain this would cause in our htaccess file.
But we are thinking in the same ballpark as you mentioned - trying to find ways to somehow limit the 404 tsunami this would cause and see if we can "take back" some of the value they took from us in link juice.
-
Yes, redirections are 100% necessary. I agree whole heartedly.
-
Surely you can block them once the contract has been ended? I don't know how the law works where you are, but in the UK if you sever a contract you are no longer bound by it. But then again, I'm not a lawyer!!! LOL I'd be earning twice as much if I was!!! I'd look into this or get your legal team (assuming you have one) to look into it for after the contract has ended.
If they're scraping, could you put a canonical tag on your pages to self canonicalise? Only just thought of this!!! Might help, if you've not already done it.
-
Hey thanks so much for the response!
And there are no stupid questions!
Before I was hired here, the company was incredibly aggressive with PPC and CSE's and spent absorbent amounts on paid traffic.
The company literally drove 2x more traffic through paid than through organic. That has changed now even though we still spend pretty aggressively. We have an excellent SEM Digital Marketing Manager that handles all paid campaigns and affiliate programs and she is run ragged on a daily basis.
I really do think it would be worth taking a look at how we can compensate with PPC on the black-hat vendor's best performing URLs and thank you so much because it is an excellent idea.
To your robot blocking question:
I would love nothing more than to insert robot text that disallows Google Bot from crawling the tree sub folders that contain all of their doorway pages. Unfortunately, they entered into a legally binding contract and this would be like an act of war against them. I actually dream about doing this to them every night so that is an awesome point you bring up!
-
Thanks so much for the response.
Your advice on having a battle plan is perfect and is something that I have had to try, try try try and once I am done trying, I try again to find more creative ways to present SEO needs, site fixes and strategies.
I even went so far to show them what their page title look like in search when they are 90 characters long and compared them to that shady gas station on an isolated highway when we could be optimizing the titles, increase CTR and add some schema to product page SERPS to make them look like Sheetz!
Full PowerPoint pictures of gas stations!
The enigma of pushing SEO when nothing is "guaranteed" but the numbers they are seeing from this black-hat vendor are.
Yesterday, digging deeper and deeper using Screaming Frog, I dug into one of this vendor's sub folders that is a giant index (They have three of these sub folders they upload to our site)
I actually found that they are literally completely copying our product pages and making exact copies. They then insert basically meta spam links on the product pages that ensures that their copies will usually always out rank our original content that we have three writers working on.
Unbelievable I know. So with your awesome advice and internal reminder on how much more I need to think outside the box with presenting, I am going to make an entire roster of this plagiarized pages and show them that if all of these copied product pages were removed, our own organic product pages would show as they are meant to.
I cannot believe vendors still can get away with this. No one monitored them or had any idea what they were doing until I was hired it is just beyond belief.
Thank you so much for the advice and inspiration.
-
Nicely put, Amelia!
PPC would definitely be a great alternative to make up any losses from organic search. And, PPC Hero is indeed a great resource as is the AdWords help center.
From a technical standpoint, one would still want to have all of those crappy vendor pages re-directed somewhere, which would be a pain to manually do but a necessary pain. If not, they would be sending a huge amount of 404 errors and that's not going to be a good sign for Google. The pages are already indexed since they are getting traffic and you'd want to send that traffic, and any links associated with that page, somewhere - ideally, in your situation, a much better (relevant) page from a user and search engine perspective.
-
What a pig awful situation to be in. I feel for you.
The previous poster has some great suggestions which I would follow.
May I also suggest that you start a PPC campaign of your own to 'pick up the slack' as you put it? Assuming the budget previously allocated to this vendor would cover it? If the vendor was using PPC as the revenue driver to these horrible UX pages, imagine how much better the conversion would be from one of your 'good' pages?
If you've never used Adowrds before, then I would look at the adwords education center for a bit (sorry I can't remember what it's called). A good site I used to use when first starting out learning Adwords is PPC Hero - they had some good tips a few years ago, and I have no reason to believe they've gone downhill! I think (and hope I don't inadvertently offend anyone here, but it's my experience) that if you can do SEO then running PPC (though time consuming) should be easy enough for you to get your head around.
I don't know if this is a stupid suggestion or not as I'm not very technical (I rely on brilliant developers in my team) but could the vendor's dodgy pages be disallowed by your robots file? Could you also remove them from the index via webmaster tools (especially if the pages are just PPC landing pages and not built for organic search, which I understand is the case from your post)? Like I say, this may be a stupid suggestion... Please go easy on me if it is!!!
Good luck - and remember, 'what doesn't kill us, makes us stronger'. I bet you're a much better SEO now than you were a year and a half ago!
-
I feel like your best plan of attack will be two sided:
1.) Education - Which is a definite struggle, but helping your higher-ups really understand WHY these practices are an issue and how it could and eventually will impact their bottom line might resonate more than just saying there are issues present (which I am sure you have been doing anyway). Perhaps reiterating the amount of revenue that is a result of natural search and how much would be lost if the site were penalized would paint a more clear picture. Having data to support your arguments is always helpful. Maybe you can even do some research and present a few summarized case studies on other sites that have been penalized and how it impacted their natural search metrics.
2.) Plan - Have a plan of attack ready. Ok, so you get rid of these pages... Now what? Preparing a very clear, step-by-step plan on what changes need to be made, what these changes will accomplish and what issues they will address, how you will make them and how long it will take, and what the expected outcome will be will help them better understand the process and how it will help save and possibly even improve revenue.
Hope this is helpful - good luck!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Old subdomains - what to do SEO-wise?
Hello, I wanted the community's advice on how to handle old subdomains. We have https://www.yoursite.org. We also have two subdomains directly related to the main website: https://www.archive.yoursite.org and https://www.blog.yoursite.org. As these pages are not actively updated, they are triggering lots and lots of errors in the site crawl (missing meta descriptions, and much much more). We do not have particular intentions of keeping them up-to-date in terms of SEO. What do you guys think is the best option of handling these? I considered de-indexing, but content of these page is still relevant and may be useful - yet it is not up to date and it will never be anymore. Many thanks in advance.
White Hat / Black Hat SEO | | e.wel0 -
A Sitemap Web page & A Sitemap in htaccess - will a website be penalised for having both?
Hi I have a sitemap url already generated by SEO Yoast in the htaccess file, and I have submitted that to the search engines. I'd already created a sitemap web page on the website, also as a helpful aid for users to see a list of all page urls. Is this a problem and could this scenario create duplicate issues or any problems with search engines? Thanks.
White Hat / Black Hat SEO | | SEOguy10 -
How does Google handle product detail page links hiden in a <noscript>tag?</noscript>
Hello, During my research of our website I uncovered that our visible links to our product detail pages (PDP) from grid/list view category-nav/search pages are <nofollowed>and being sent through a click tracking redirect with the (PDP) appended as a URL query string. But included with each PDP link is a <noscript>tag containing the actual PDP link. When I confronted our 3rd party e-commerce category-nav/search provider about this approach here is the response I recieved:</p> <p style="padding-left: 30px;">The purpose of these links is to firstly allow us to reliably log the click and then secondly redirect the visitor to the target PDP.<br /> In addition to the visible links there is also an "invisible link" inside the no script tag. The noscript tag prevents showing of the a tag by normal browsers but is found and executed by bots during crawling of the page.<br /> Here a link to a blog post where an SEO proved this year that the noscript tag is not ignored by bots: <a href="http://www.theseotailor.com.au/blog/hiding-keywords-noscript-seo-experiment/" target="_blank">http://www.theseotailor.com.au/blog/hiding-keywords-noscript-seo-experiment/<br /> </a> <br /> So the visible links are not obfuscating the PDP URL they have it encoded as it otherwise cannot be passed along as a URL query string. The plain PDP URL is part of the noscript tag ensuring discover-ability of PDPs by bots.</p> <p>Does anyone have anything in addition to this one blog post, to substantiate the claim that hiding our links in a <noscript> tag are in fact within the SEO Best Practice standards set by Google, Bing, etc...? </p> <p>Do you think that this method skirts the fine line of grey hat tactics? Will google/bing eventually penalize us for this?</p> <p>Does anyone have a better suggestion on how our 3rd party provider could track those clicks without using a URL redirect & hiding the actual PDP link?</p> <p>All insights are welcome...Thanks!</p> <p>Jordan K.</p></noscript></nofollowed>
White Hat / Black Hat SEO | | eImprovement-SEO0 -
Best practice to preserve the link juice to internal pages from expired domain?
This question relates to setting up an expired domain, that already has quality links, including deep links to internal pages. Since the new site structure will be different, what's the best practice to preserve the link juice to these internal pages? Export all the internal pages linked to using majestic Seo/ ahrefs etc, and set these pages previously linked to? Or 301 redirect these pages to home page? I heard there's a Wordpress plugin that 301 redirects all the 404 errors successfully preserving all the potential link juice.
White Hat / Black Hat SEO | | adorninvitations0 -
Redirecting old domains for SEO ranking?
It's been a while since I read anything seriously out of the box on SEO but I thought I would see what others thought of the bold assertions made in this article. Most of it revolves around buying expired domains and using a 301 to point them, and their juice, to new sites. This guy makes a living doing this so he has to know a bit more than the average Joe but I'm wondering where the other shoe is and when it drops.
White Hat / Black Hat SEO | | Highland0 -
Pages linked with Spam been 301 redirected to 404\. Is it ok
Pl suggest, some pages having some spam links pointed to those pages are been redirected to 404 error page (through 301 redirect) - as removing them manually was not possible due to part of core component of cms and many other coding issue, the only way as advised by developer was making 301 redirect to 404 page. Does by redirecting these pages to 404 page using 301 redirect, will nullify all negative or spam links pointing to them and eventually will remove the resulting spam impact on the site too. Many Thanks
White Hat / Black Hat SEO | | Modi0 -
Changes to SEO with disavow?
Has the game changed a lot with the disavow tool I can see people still saying check out what our competitors are doing but with just going through a disavow myself how do you actually know what the correct link diversity is as 0 - 100% of the links could be disavowed. Also could a competitor not just buy a load of spammy links and disavow them to mask there real links. (I know in my backlinks on 150 are good and the rest is disavowed crap)
White Hat / Black Hat SEO | | BobAnderson0 -
Can a Page Title be all UPPER CASE?
My clients wants to use UPPER CASE for all his page titles. Is this okay? Does Google react badly to this?
White Hat / Black Hat SEO | | petewinter0