How to go about removing bad/irrelevent links?
-
We have been made aware of a series of irrelevant links on some rather dodgy sites.
- http://www.designerdogstop.com/level-static/http://www.bestfirepits.net/some-of-the-best-vacations-for-families/2010/05/05/
- http://whatcigarsdoismoke.com/cigar-lighters/cigarette-cigar-2/
- http://dollfuss.org/build-bear-hawaii/
Absolute rubbish im sure you will agree. These links must surely be causing our link profile some damage.
They are currently wordpress sites with no means of contacting the authors. What ways and means are there of removing these negative pages and links?
Cheers all, any help appreciated.
-
Thanks Daniel,
Lots of additional features & improvements still in the works - updates @rmoov
Sha
-
Nice! Checking it out right now. Good luck with the project ! ~^DH
-
Hi Tim,
We just opened free Beta during this last weekend on a link removal management service that we developed.
The free trial will allow you to run a campaign including the domains you mentioned in your post.
The tool allows you to drop in a list of URLs, pull contact information from ICANN (the whois data mentioned above) for each of the domains, customize emails, send follow-ups, receive notification of cleanups from webmasters etc.
Hope it helps,
Sha
-
There's no quick route for SEO. Work slow, and do your work properly. I think that ipositions' suggesting is by far the best one - look-up contact details, contact them and request to remove the link. Patience helps
-
I do agree that not spending too much time is critical, as it can become hugely time consuming and inefficient. But getting rid of a few is also a help.
Being so time consuming was the main reason in my initial post, to see if there were any quicker options available to web masters, e.g. sending google a list of links that you wish for it to disregard due to not being generated by yourself... unlikely I know... :o(
Thanks for the response.
-
Have you looked for the email address of the webmaster using whois.net? That's the first thing we do when contact details are not provided on the site.
I've seen a lot of people here going on about removing bad links and where possible I agree that it should be done. However a line must be drawn where getting a link removed becomes too time consuming to be worth the effort. That time would be better spent creating content, reaching out to popular relevant blogs etc and create enough good links that offset the bad ones.
Hope this helps
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Yelp (recrawl Google/Bing)
If Google and Bing show an older version of a site's Yelp rating in the search results, what options are there to help ensure Google and Bing recrawl the Yelp page? Additionally, it appears third-party sites such as MapQuest show Yelp ratings and appear in Google search results; is it possible to request MapQuest to recrawl Yelp and then ask Google to recrawl MapQuest? Any advice would be much appreciated!
Industry News | | Mack_1 -
What is the most frustrating/challenging part about sending marketing emails?
I'm in the process of gathering research into the challenges most commonly faced by marketers when executing email campaigns. What are yours? And what role do your play in the process? (Job title?) Looking forward to a spirited discussion!
Industry News | | johnbonini0 -
Pre-launch market/industry research
I am getting ready to start a web based marketing firm, with a strong focus on local search and b2b organic search engine optimization. However, I want to pick a specific b2b market to focus on when the site goes live. How do I go about conducting this market research when i do not have any visitors to my site yet. My goal is to spot a market/industry that has a gap where my service will be in demand? All of the popular guides and blogs focus on market research that takes place after you launch your website or after you already have an audience to analyze. Any advice or suggestions on conducting market/industry research without having a audience yet would be greatly appreciated!
Industry News | | ranch1130 -
SEO Guidelines for Partner Links
We decided to follow an aggressive strategy with cross linking. We have cross linked all all the 6 different brands just above the header of a News Portal. You could see what i mean here www.in.gr
Industry News | | DOLdatacenter
Do you think that i should enable the nofollow so as not to get penalized by Google. We tend to use the same cross network linking scheme (in the header) for all branded sites. Google’s guidance on link schemes doesn’t specifically address these types of partner links, which don’t really fit the mold for “excessive link exchanging.” Thoughts appreciated.0 -
How do I predict quality of inbound link before using Disavow links tool?
I am working on Ecommerce website and getting issues with bad inbound links. I am quite excited to use Disavow links tools for clean up bad inbound links and sustain my performance on Google. We have 1,70,000+ inbound links from 1000+ unique root domains. But, I have found maximum root domains with low quality content and structure. Honestly, I don't want to capture inbound links from such websites who are not active with website and publishing. I am quite excited to use Disavow links tool. But, How do I predict quality of inbound links or root domains before using it? Is there any specific criteria to grade quality of inbound links?
Industry News | | CommercePundit0 -
Concerned about where links are coming from
Hi I am hoping someone can help me with this. I run a small seo company for a few select companies in the luxury markets. Due to certain circumstances I need to outsource a lot of link building until my in house link builders can return to work. After a phone conversation with an SEO company which offer white label service (quite a large UK company but will remain nameless) I became extremely concerned that they are supply paid links to there customers. I will explain what i mean and would appreciate advice on whether I should stay clear or not. They claim to have access to over 1000 websites in many different niches and can get a whole host of links from contextual (which they said would give 3 links in 300 words which rang the first alarm bell) then said the other come from primary theme links and hybrid links (oh which neither they would properly explain what they are). My main concern is they said the links are made and need to be "maintained" by a yearly fee so they are not removed. When I asked if they surely count as paid links I got a usual boiler room style sales pitch. My instincts tell me to stay away but maybe I am just reading it all wrong. They claim to already white label to over 300 other SEO companies, but it all sounds a bit risky to me after Googles tough stance on this! Thanks Paul
Industry News | | hanv0 -
LinkSmart Raises $4.7M to Dynamically Change Links
LinkSmart is a startup trying to put a new spin on link management by providing a platform to dynamically change links. They just raised $4.7M, and TechCrunch posted a quick blurb on them here that is worth reading. In short, the application will allow a site owner to change all of the text links on their site in real-time. The idea is that this will help improve click-through rates, optimize traffic, etc. How do you guys see this from an SEO perspective? It seems a bit scary from an indexing standpoint if links are never consistent. Also, the flow of link juice will be inconsistent. Any thoughts?
Industry News | | tailwindcreative0 -
What is the best method for getting pure Javascript/Ajax pages Indeded by Google for SEO?
I am in the process of researching this further, and wanted to share some of what I have found below. Anyone who can confirm or deny these assumptions or add some insight would be appreciated. Option: 1 If you're starting from scratch, a good approach is to build your site's structure and navigation using only HTML. Then, once you have the site's pages, links, and content in place, you can spice up the appearance and interface with AJAX. Googlebot will be happy looking at the HTML, while users with modern browsers can enjoy your AJAX bonuses. You can use Hijax to help ajax and html links coexist. You can use Meta NoFollow tags etc to prevent the crawlers from accessing the javascript versions of the page. Currently, webmasters create a "parallel universe" of content. Users of JavaScript-enabled browsers will see content that is created dynamically, whereas users of non-JavaScript-enabled browsers as well as crawlers will see content that is static and created offline. In current practice, "progressive enhancement" in the form of Hijax-links are often used. Option: 2
Industry News | | webbroi
In order to make your AJAX application crawlable, your site needs to abide by a new agreement. This agreement rests on the following: The site adopts the AJAX crawling scheme. For each URL that has dynamically produced content, your server provides an HTML snapshot, which is the content a user (with a browser) sees. Often, such URLs will be AJAX URLs, that is, URLs containing a hash fragment, for example www.example.com/index.html#key=value, where #key=value is the hash fragment. An HTML snapshot is all the content that appears on the page after the JavaScript has been executed. The search engine indexes the HTML snapshot and serves your original AJAX URLs in search results. In order to make this work, the application must use a specific syntax in the AJAX URLs (let's call them "pretty URLs;" you'll see why in the following sections). The search engine crawler will temporarily modify these "pretty URLs" into "ugly URLs" and request those from your server. This request of an "ugly URL" indicates to the server that it should not return the regular web page it would give to a browser, but instead an HTML snapshot. When the crawler has obtained the content for the modified ugly URL, it indexes its content, then displays the original pretty URL in the search results. In other words, end users will always see the pretty URL containing a hash fragment. The following diagram summarizes the agreement:
See more in the....... Getting Started Guide. Make sure you avoid this:
http://www.google.com/support/webmasters/bin/answer.py?answer=66355
Here is a few example Pages that have mostly Javascrip/AJAX : http://catchfree.com/listen-to-music#&tab=top-free-apps-tab https://www.pivotaltracker.com/public_projects This is what the spiders see: view-source:http://catchfree.com/listen-to-music#&tab=top-free-apps-tab This is the best resources I have found regarding Google and Javascript http://code.google.com/web/ajaxcrawling/ - This is step by step instructions.
http://www.google.com/support/webmasters/bin/answer.py?answer=81766
http://www.seomoz.org/blog/how-to-allow-google-to-crawl-ajax-content
Some additional Resources: http://googlewebmastercentral.blogspot.com/2009/10/proposal-for-making-ajax-crawlable.html
http://www.seomoz.org/blog/how-to-allow-google-to-crawl-ajax-content
http://www.google.com/support/webmasters/bin/answer.py?answer=357690