Google launches their Disallow Tool
-
-
Hi Irving,
Today i have Compiled a comprehensive backlink report for one of my hotel client (http://www.fairfieldinnhotelcedarrapids.com/).I have identified all bad links and create one .txt file and upload via google disallow tool. now my question is, How do i know that all dead links has been removed? Is there any way to know status?
Thanks
-
If you Disavowed "good" links to your site, your rankings may be lowered as a result.
If you Disavowed "bad" links, then your Penguin issue could be reduced or resolved.
The best course of action is to allow a trained SEO professional examine each linking domain to determine if the links violate Google's Guidelines and only Disavow the links which do violate those guidelines. Also, the Disavow tool should not be used until after every possible action has been taken to remove the link. Google is quite clear on this topic. If you use the Disavow tool without "significantly" reducing the manipulative links to your site, it likely will not help.
-
I admit it... I panicked and disavowed a ton of domains that I didn't recognize as good links and I went from mid-page-2 to lower-page-3. So...my question is: what happens if I submit a new disavow file that has only a fraction of the links. Or, better yet, what if I delete the disavow file altogether with a file that has a comment saying Oops, sorry, we didn't know what we were doing with this tool and respectfully request to undo our mess".
A very good SEO told me to not chase after my disavow list, and I get what he's saying, but it's hard to not remember the days pre-disavow when I was at least on page 2 ... I was on page 1 for 14 years since 1998 until penguin hit! (I sell bean bag chairs and am speaking of serps for "bean bag chairs").
www.ahhprods.com in case anyone is curious
Thanks!
-
So far my webmaster response is about 10%, so you have no idea how much this tool can relieve some pain after so many attempts.
To the best of my knowledge, the Disavow Tool will have absolutely no impact on your success rate. It seems designed to help ensure webmasters who have a manual penalty lifted are not affected by Penguin.
-
This is gonna be interesting how everything works out.
I've sent a reconsideration request a day or two prior to the release. I will see how that request goes and then update the spreadsheet with new links I've removed as well as using disavow.
So far my webmaster response is about 10%, so you have no idea how much this tool can relieve some pain after so many attempts.
-
The damage was done before i came on the picture and there is no stats from before. the site rangs below other sites owned by the same clent with much weaker linking profiles. As i mentioend the site has the best of links, this is why i am prepared to get rid of any links that look even the slightest doggy. The site should rank number 1 when you look at the competition
I just dont have the time or budget to try to ask for removal, a attemp was made long ago with no sucess.
-
Very busy with a new project out of Arizona.
i have been following Mitt closely. He might just get over the line.
-
-
Thanks... Should have read the YouTube description.
-
This says to me that your site was spared an across the board penalty, but your rankings for specific keywords that have been overused in your anchor text have been suppressed. I would look at your incoming link anchor texts and see which one(s) you are no longer ranking for.
-
Excellent sir.
We will know pretty soon how everything shakes out once people start reporting back, but my suggestion would be that if you have a site that is not penalized you should NOT use this tool as an effort to try and clean up any spammy back links and clean up your good-to-bad backlink ratio. The reason is, this is a tool to be used as a last effort in trying to come back from a penalty when there are some links you tried to remove but simply cannot.
Sending this report will put eyeballs on your site and bring unnecessary attention to your site. Why ask Google to review your backlink profile and look at the nastiest links pointing to your site if your site is currently healthy.
An exception to this rule I think would be if you notice you are clearly under a negative SEO attack. Then it would make sense to be proactive.
-
Very solid analysis Ryan, good stuff.
-
Long time comming and quiite a messy interface. why they could not do somthing like Bing did with there tool is a mystery.
I have a client with a unatural link warning, saying "for this specific incident we are taking very targeted action on the unnatural links instead of your site as a whole"
to me this sounds like these links have been discounted anyhow and that the site is not punished, and maybe no need to do anything, but then goes on to say "If you are able to remove any of the links, please submit a reconsideration request, including the actions that you took."
so that makes me think i do need to do something. not very clear.
This client has a lot of very good links from CNN, NYT and a host of others, but partisipated in a link wheel. The blogs in this link wheel are real blogs rather than your obvious mass made for links blogs and makes it hard to identify what are what. i am thinking to disavow anything thats could be doggy, he has such good links I think it better to have a few false positives rather then leave any bad links in the profile.
Back to my first point, i was hoping for a click and job done approch link in BWMT.
-
Nice write-up Ryan, thanks.
Looks like an aggressive tool, I can see a lot of Webmasters running into trouble with this one.
If you contact a blog to get a link removed and then realise after actually I have made a mistake, or you login to a directory and remove it yourself and realise after you have made a mistake you can work on getting it back.
I get the feeling if you don't go through the process Ryan put down and you make a mistake with the Disallow Tool you wont be able to get those links back.
-
You are so right Ryan! This tool is not a shortcut at all. I fear that a lot of webmasters who have an unnatural links warning are going to jump straight to the disavow tool and ignore the actual reconsideration request process. As Matt says in the video, you still need to make a thorough attempt at trying to get the links removed on your own in order to have a manual penalty revoked.
-
Thanks for opening this discussion Irving. I have had calls from clients today regarding this "change" and it seems many site owners are simply caught up in the idea without realizing the true impact of this change. Resolving a manual Google penalty for manipulative links involves 4 steps prior to the release of this tool:
1. Compile a comprehensive backlink report. Many sites which suffer from a manipulative link penalty are absolutely doomed to have their Reconsideration Requests declined before they are even submitted. Why? Because they have not captured all the links to their site. You cannot rely on any single tool or even 2 tool combination. For each client I work with we compile a list of every known backlink to their site. How? By combining Google + Bing + OSE + Majestic + AHREFs data. Each data source offers links the others do not seem to find.
2. Properly identify all the manipulative links to the target site. Once again, many site owners repeatedly fail their Reconsideration Request and have no real chance at success because they try to take the easy way out. Attempting to replace real effort with fake work is what caused the penalty in the first place.
a. A thorough understanding of the difference between an organic link and manipulative link is required. In short, you must calibrate your understanding of links to match Google. How do you view free directory links? The reality is 99%+ of them are manipulative. How about press releases? Do you think most press releases are organic links? When site owners pay another company to publish articles they wrote with links back to their site, does that sound natural to you?
b. How about broken links? Can you use an automatic link checker and then if the link is not on the URL simply cross it off the list? In a significant percent of cases the link has simply moved to another page on the linking site. Some sites have very dynamic link structures where one day a URL is at ?page=20 and the next it is at ?page=21. Other sites make URL changes over time. You must search each site using their search widget and a Google site: search before assuming the site's link is gone.
c. Is the link marked NoFollow? You need to keep searching the page to ensure there are not other followed links on the same page.
The above are just some examples of gaps in the process of many who attempt to resolve this type of penalty. The disavow tool's introduction does not impact this step.
3. Webmaster Outreach. Once you have a comprehensive list of all known links to your site and have properly identified all the manipulative links, there is a need to contact every site on the list. Another common issue is those attempting to resolving a manipulative link issue give up far too easy. Site owners can be contacted via their WHOIS email address, the email address on their site AND the contact form on their site. You can call them, send a letter and chase them down on social networks. This type of sincere effort can lead to 50%+ reduction in links to your site.
Once sincere and comprehensive efforts have been made to remove the links, Google can clearly tell because there will be a "significant" reduction in manipulative links. At that point, THEN the Disavow tool can be used.
4. Filing a thoroughly documented Reconsideration Request. Three days later, the Reconsideration Request can be submitted.
So the introduction of this tool actually did not reduce any step in the process at all. Matt clearly outlined Google's expectation the tool is only used after a webmaster outreach campaign has been completed. If you expect to be able to simply submit a list of links without webmaster outreach, you are likely going to be disappointed.
Watch the first 2 minutes of the video a few times. Matt clearly says "...when you have contacted each webmaster multiple times....and there are only a small fraction of links left....that is when you can use the tool."
-
Nice summary at SEWatch:
http://searchenginewatch.com/article/2217602/Google-Disavow-Links-Tool-Now-Available
I'm curious about the "Most sites shouldn’t use this tool,” Cutts said. “Use caution." caveat. I've basically got only one client out of many that I'd even need to consider using this for. But I can't help but imagine hyper paranoid SEOs trying to massage their link profile down to the last drop of relevance. My gut feel is that this is a 'last resort' tool, and not a 'everyday SEO' tool.
-
Based on what Matt said, it sounds like Google only wants you to use the tool for links that you've tried to remove manually but couldn't. My guess is they may ignore your disavowals if you rely too much on the tool.
-
The links are in the YouTube video description.
"Access the feature here:
https://www.google.com/webmasters/tools/disavow-links-main" -
Cutt's did not indicate how to download the tool. Did I miss that?
-
Thanks for the heads up. Just watched the video.
-
Really looking forward to this tool... But Joeys question is Really important. Does anyone have suggestions?
-
Awesome tool. How can I tell which links to my site I should disavow? We get a bunch of random links per day that look spammy, but how can I tell for sure that removing them will help rather than hurt?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google says Geolocation Redirects Are Okay - is this really ok ?
Our aim is to send a user from https://abc.com/en/us to** https://abc..com/en/uk/ **if they came to our US English site from the UK So we came across this document - https://webmasters.googleblog.com/2014/05/creating-right-homepage-for-your.html We are planning to follow this in our international website based on the article by google : automatically serve the appropriate HTML content to your users depending on their location and language settings. You will either do that by using server-side 302 redirects or by dynamically serving the right HTML content. Will there be any ranking issues/ penalty issue because of following this or because of 302 redirects ? **Another article - **https://www.seroundtable.com/google-geolocation-redirects-are-okay-26933.html
White Hat / Black Hat SEO | | NortonSupportSEO0 -
How does Google handle product detail page links hiden in a <noscript>tag?</noscript>
Hello, During my research of our website I uncovered that our visible links to our product detail pages (PDP) from grid/list view category-nav/search pages are <nofollowed>and being sent through a click tracking redirect with the (PDP) appended as a URL query string. But included with each PDP link is a <noscript>tag containing the actual PDP link. When I confronted our 3rd party e-commerce category-nav/search provider about this approach here is the response I recieved:</p> <p style="padding-left: 30px;">The purpose of these links is to firstly allow us to reliably log the click and then secondly redirect the visitor to the target PDP.<br /> In addition to the visible links there is also an "invisible link" inside the no script tag. The noscript tag prevents showing of the a tag by normal browsers but is found and executed by bots during crawling of the page.<br /> Here a link to a blog post where an SEO proved this year that the noscript tag is not ignored by bots: <a href="http://www.theseotailor.com.au/blog/hiding-keywords-noscript-seo-experiment/" target="_blank">http://www.theseotailor.com.au/blog/hiding-keywords-noscript-seo-experiment/<br /> </a> <br /> So the visible links are not obfuscating the PDP URL they have it encoded as it otherwise cannot be passed along as a URL query string. The plain PDP URL is part of the noscript tag ensuring discover-ability of PDPs by bots.</p> <p>Does anyone have anything in addition to this one blog post, to substantiate the claim that hiding our links in a <noscript> tag are in fact within the SEO Best Practice standards set by Google, Bing, etc...? </p> <p>Do you think that this method skirts the fine line of grey hat tactics? Will google/bing eventually penalize us for this?</p> <p>Does anyone have a better suggestion on how our 3rd party provider could track those clicks without using a URL redirect & hiding the actual PDP link?</p> <p>All insights are welcome...Thanks!</p> <p>Jordan K.</p></noscript></nofollowed>
White Hat / Black Hat SEO | | eImprovement-SEO0 -
Google Disavow and Penalty lifted please help?
We disavowed 80% of our backlink profile due to our last SEO building cheap nasty links and filed a reconsideration requested (we had the Google Webmaster Tools notice of detected unnatural links to http://www.xxx.co.uk penalty for a year from the 24<sup>th</sup> march 2012 but thought it best to clean up before round 2 – even though we had no real penalty and we dd some decent link building that moved us up). We then received a successful penalty lifted note (on the 22<sup>nd</sup> of May 2013) but our rankings dropped (due to the crap links propping us up) since then we have built a fair few high quality links but our rankings do not seem to be moving much if at all (7 weeks clear now). has anyone had any experience with the above (are we in a sandbox type situation). Thank you for your time Thanks Bob
White Hat / Black Hat SEO | | BobAnderson0 -
If Google Authorship is used for every page of your website, will it be penalized?
Hey all, I've noticed a lot of companies will implement Google Authorship on all pages of their website, ie landing pages, home pages, sub pages. I'm wondering if this will be penalized as it isn't a typical authored piece of content, like blogs, articles, press releases etc. I'm curious as I'm going to setup Google Authorship and I don't want it to be setup incorrectly for the future. Is it okay to tie each page (home page, sub pages) and not just actual authored content (blogs, articles, press releases) or will it get penalized if that occurs? Thanks and much appreciated!
White Hat / Black Hat SEO | | MonsterWeb280 -
Manage two domains with Webmaster tools
Dear all, I want to create an additional domain in order to: Rank better for a very specific keyword with an exact match domain (I already ask about that, but I did not have my ideas clear at the time); Offer to the user usefull infomation about the topic, without duplicating the content I have in my main domain, just additional and very specific information; Use this domain as landing page, offering a tutorial on "how to" use a specific section of my main domain, including a video tutorial; Link to the related section of my main domain. So, the main idea is, if an user type in google "this specific keyword of ours", they will have in the results "thisspecifickeywordofus.es", they will click and go to the site, where they will find unique and specific information, complementing what I have in my mainsite, and showing how to use my site, so trying to use it for conversion. I want to do only white hat SEO, so first at all, I would like to ask you if you think it is a good idea. The keyword is difficult to rank for, and if I can take advantage of this exact match domain (even if it is nowdays no more so big an advantage), would be great. Second, do you see any problem in managing different domains from the same google account? Newbee question, sorry. Thanks in advance for your help, Daniel
White Hat / Black Hat SEO | | te_c0 -
Google Sitemaps & punishment for bad URLS?
Hoping y'all have some input here. This is along story, but I'll boil it down: Site X bought the url of Site Y. 301 redirects were added to direct traffic (and help transfer linkjuice) from urls in Site X to relevant urls in Site Y, but 2 days before a "change of address" notice was submitted in Google Webmaster Tools, an auto-generating sitemap somehow applied urls from Site Y to the sitemap of Site X, so essentially the sitemap contained urls that were not the url of Site X. Is there any documentation out there that Google would punish Site X for having essentially unrelated urls in its sitemap by downgrading organic search rankings because it may view that mistake as black hat (or otherwise evil) tactics? I suspect this because the site continues to rank well organically in Yahoo & Bing, yet is nonexistent on Google suddenly. Thoughts?
White Hat / Black Hat SEO | | RUNNERagency0 -
Does Google+ make a huge difference?
I run a website that's been ranked well for good keywords related to our business for some time. It was founded back in 2007 and has been there a while. Recently a new site has popped up that ranks brilliantly for everything. It's a new site, and the only redeeming factor I can see is that it has an AddThis box showing the Facebook Likes and Google Plus Ones, and they are around 400 Facebook Likes and 80 Google+ (for every page that ranks). Any other pages on their site which doesn't have any Facebook likes or Google Plus Ones, they don't rank. Our site doesn't have any likes or pluses. Is this making the difference? I stress that other than this our sites are very similar, other than the fact we've been around over 5 years.
White Hat / Black Hat SEO | | freebetinfo0 -
Switching prices for google base
We would like to be able to submit lower prices to google than we do to other sources. How i see it working is that at the end of each url we submit to google base there is a tracking code (source=googlebase). When a user visits the site via one of these urls we would knock 10% of the price of that item and store the item in a cookie to ensure that the price of that item, for that user would stay at the low price for 24 hours. My question is whether google would have a problem with us doing this? The second part of my question is whether they check the full url including the query strings? If theyt just checked the canocial URL they would see a price thats 10% higher than the one we submitted to base - which, of course - would be bad
White Hat / Black Hat SEO | | supermarketonline0