Discovering broken links to my site
-
How do I find out if other websites are trying to link to my website using a misspelled URL or broken link?
-
One extra resource is this article
http://moz.com/blog/set-it-and-forget-it-seo-chasing-the-elusive-passive-seo-dream
It gives a little more info in researching broken links from other sites, and even gives a great explanation of using the levenshtein distance to automatically correct incoming broken links on the fly.
-
:))
-
Beat me to the punch as I was writing up my response. Thumbs up to you for a good answer that was similar to mine.
-
Depends on the way in which they misspelled things. If they got your domain name correct but messed something up after the slash (i.e. mysite.com/TyposHappne) then you will likely see a notice in Google Webmaster Tools of a 404'd page. At which point you could either ask the webmaster at the other site to fix the link or you could 301 the 404's page to the correct URL. If the incorrect link spells your domain name wrong... well then there's really no easy way to find it. You could potentially run a crawl test on a website that you know does link to you on other pages to see if the test brings up notices of broken links and then check if those broken links were meant for your site. Beyond that I don't believe you can do anything except hope the site administrator or their SEO runs their own tests to check for broken links and then fixes them accordingly.
-
One of the best way to check links that comes to your site is to use backlink checker software (open site explorer). If the main url is mis-spelt i don't think you can find out using the software as the url wouldn't point to your website at all. If you requested the link your self from a website, you can only find out by visiting the website and checking the link.
To check the broken links, I use broken link checker software and i do redirect or fix the links so if there are any links coming through to those pages it'll be fixed or redirected. I use brokenlinkcheck.com or screaming frog to check broken links.
I hope that helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Can you track two Google Analytics Accounts on one site?
If you have a site that had an old analytics account and then implemented a new one is it possible to run tracking code that records to both accounts without causing your site or data issues? We are doing this so we don't loose data at any point - ideally it wouldn't have been split between the two but making one redundant isn't an option. Ideally we would have merged the data from both accounts and had one - however the research we have done points to this not being a possibility - unless one of you guys knows different? It would be great if anyone has experience on any this.. Thanks
Reporting & Analytics | | ChrisAllbones0 -
Getting google impressions for a site not in the index...
Hi all Wondering if i could pick the brains of those wise than myself... my client has an https website with tons of pages indexed and all ranking well, however somehow they managed to also set their server up so that non https versions of the pages were getting indexed and thus we had the same page indexed twice in the engine but on slightly different urls (it uses a cms so all the internal links are relative too). The non https is mainly used as a dev testing environment. Upon seeing this we did a google remove request in WMT, and added noindex in the robots and that saw the index pages drop over night. See image 1. However, the site still appears to getting return for a couple of 100 searches a day! The main site gets about 25,000 impressions so it's way down but i'm puzzled as to how a site which has been blocked can appear for that many searches and if we are still liable for duplicate content issues. Any thoughts are most welcome. Sorry, I am unable to share the site name i'm afraid. Client is very strict on this. Thanks, Carl image1.png
Reporting & Analytics | | carl_daedricdigital0 -
Links From Public Info
Hi guys, I'm conducting a link analysis for one of my websites and I have found a few potentially damaging links. However, there are some people who I have contacted about removing links who have said that they will not remove my link because they are using data that is publicly available. Is there anything I can do to get those links removed rather than diavow? Thanks.
Reporting & Analytics | | AAttias0 -
Google Analytics VS target="_blank" internal links: How much wrong is it?
I am working on an e-commerce website, and our CEO is sure that having target="_blank" in internal search result is boosting the conversion (not sure, but it's not an issue at the moment). The problem is that Google Analytics sees all URLs visited from search results as entrances/direct visits, hence the Booking Funnel Tracking does not work as it was supposed to. Is there any way to recover the tracking? Or we shall get the rid of target="_blank" attribute?
Reporting & Analytics | | apartmentGin0 -
Google Webmaster Tools - When will the links go away!?
About 9 months back we thought having an extremely reputable company build our client some local citations would be a good idea. You definitely know this citation company, but I'll leave names out. Regardless, it's our mistake to cut corners. Google Webmaster Tools quickly picked up these new citations and added them to the links section. One of these citation spawned a complete mess of about 60K+ links on their network of sites through ridiculous subdomains of every state in the country and so many other domain variations. We immediately went into remove mode and had the site's webmaster take down the bad links from their site. This process took about a month for outreach. The bad links (60K+) have not been on the spam site for well over 6 months but GWT still shows them in the "links to your site" section. Majestic, Bing, and OSE only displayed the bad links for a brief time. Why is webmaster tools still showing these links after 6+ months? We typically see GWT update about every 2 weeks, a month tops. Any ideas? Could a changed robots.txt on the bad site prevent Google from updating the links displayed in GWT? We have submitted to disavow, but Google replied with "no manual penalty". We even blasted the bad site with Fiverr links, in hopes that Google would re-crawl them. No luck with anything we do. We have patiently waited for way too long. The rankings for this site got crushed on Google after these citations. How do we fix this? Should we worry about this? Any advice would really help. Thanks so much in advance.
Reporting & Analytics | | zadro0 -
Any harm and why the differences - multiple versions of same site in WMT
In Google Webmaster Tools we have set up: ourdomain.co.nz
Reporting & Analytics | | zingseo
ourdomain.co.uk
ourdomain.com
ourdomain.com.au
www.ourdomain.co.nz
www.ourdomain.co.uk
www.ourdomain.com
www.ourdomain.com.au
https://www.ourdomain.co.nz
https://www.ourdomain.co.uk
https://www.ourdomain.com
https://www.ourdomain.com.au As you can imagine, this gets confusing and hard to manage. We are wondering whether having all these domains set up in WMT could be doing any damage? Here http://support.google.com/webmasters/bin/answer.py?hl=en&answer=44231 it says: "If you see a message that your site is not indexed, it may be because it is indexed under a different domain. For example, if you receive a message that http://example.com is not indexed, make sure that you've also added http://www.example.com to your account (or vice versa), and check the data for that site." The above quote suggests that there is no harm in having several versions of a site set up in WMT, however the article then goes on to say: "Once you tell us your preferred domain name, we use that information for all future crawls of your site and indexing refreshes. For instance, if you specify your preferred domain as http://www.example.com and we find a link to your site that is formatted as http://example.com, we follow that link as http://www.example.com instead." This suggests that having multiple versions of the site loaded in WMT may cause Google to continue crawling multiple versions instead of only crawling the desired versions (https://www.ourdomain.com + .co.nz, .co.uk, .com.au). However, even if Google does crawl any URLs on the non https versions of the site (ie ourdomain.com or www.ourdomain.com), these 301 to https://www.ourdomain.com anyway... so shouldn't that mean that google effectively can not crawl any non https://www versions (if it tries to they redirect)? If that was the case, you'd expect that the ourdomain.com and www.ourdomain.com versions would show no pages indexed in WMT, however the oposite is true. The ourdomain.com and www.ourdomain.com versions have plenty of pages indexed but the https versions have no data under Index Status section of WMT, but rather have this message instead: Data for https://www.ourdomain.com/ is not available. Please try a site with http:// protocol: http://www.ourdomain.com/. This is a problem as it means that we can't delete these profiles from our WMT account. Any thoughts on the above would be welcome. As an aside, it seems like WMT is picking up on the 301 redirects from all ourdomain.com or www.ourdomain.com domains at least with links - No ourdomain.com or www.ourdomain.com URLs are registering any links in WMT, suggesting that Google is seeing all links pointing to URLs on these domains as 301ing to https://www.ourdomain.com ... which is good, but again means we now can't delete https://www.ourdomain.com either, so we are stuck with 12 profiles in WMT... what a pain.... Thanks for taking the time to read the above, quite complicated, sorry!! Would love any thoughts...0 -
Backlinks vs Incoming links
Hi, I've been getting stuck into some SEO analysis for a company I work for and I am a little confused. I've tried a search to get an answer but this has ended up being more confusing. The company has been around for decades and their website since 1996. I read everywhere about 'backlinks'. My SEO toolbar shows ZERO backlinks in Google but 218 in Bing. Google Webmaster tools shows nearly 2,000 incoming links from other sites. Is a backlink the same as an incoming link? Why is this tool showing zero? I am even getting email from SEO spammers saying my backlink count is ZERO. But I can see links everywhere I look to the site. Also, on the link analysis tool with SEOMOZ a competitor is showing 3000 external showing links with 250,000 total links. My site is showing 50 and 470 respectively. I have spent the best part of two years getting the site listed in industry related directories. We have paid for entries in Yahoo and some other high (PR) -ranking directories. Prior to me there was someone else adding the site to directories and getting incoming links from industry related sites. So this has been going on a while. Why are the backlings showing as zero but links from external sites showing over 1800? Thanks TT
Reporting & Analytics | | TheTub0 -
Open Site Explorer Discrepency
Hi, I'm just starting to look into this tool as a result of the need to analyse some work that is being carried out on my site by an external SEO company that is charging absolute top dollar. I'm sure it's a simple answer, but looking at the attached image you will see that the total links is shown as 219 which is described as links from all sources. However the report that is on the page shows only 44 links which would appear to be a report based on the same criteria. What is the difference in the numbers please? Simon siteExplorer.jpg
Reporting & Analytics | | simonphumphries0