Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Best free tool to check internal broken links
-
Question says it all I guess. What would your recommend as the best free tool to check internal broken links?
-
[Spammy comment removed by forum moderator.]
-
How awesome is that Screaming Frog tool?
-
Hey Kevin,
If you download the CSV report, we should list the referring page in the CSV. I know it'd be better to have it on the on-screen report, and it is on our wish list, but you should be able to see the info in an export.
-
Hi Kevin
Thanks - yes, although I am a Moz Associate, out of full transparency I agree with that you are saying. It is not a tool I specifically use with Moz, and do prefer Screaming Frog for the exact functions you are referring to. Screaming Frog will tell you what pages the bad links are on.
You can definitely pass this on to the product team as a feature request on this page. Personally I love tools like their analytics, keyword difficulty, rank tracker, open site explorer, but I agree from this aspect the crawler is not the strongest, and I would suggest supplementing with Screaming Frog (as I do).
-Dan
-
Dan, although I love me some MOZ the crawl diagnostics kinda suck... all they do is report errors but they don't give any insight on where the bad link is originated... it merely shows the page which is 404ed, which is a BIG FAT WASTE OF TIME.
You guys should know that we need to know WHICH PAGE THE BAD LINK IS ON, and furthermore WHAT LINE OF CODE HAS THE BAD LINK. Who cares about the broken page!?!?!?!?!?
-
Yes when you set up a campaign with Moz Analytics, Moz will crawl your website and return a whole report of suggestions. It will start a sample crawl which gets returned to you pretty quickly, and a full crawl a little bit thereafter. It will report things to you like 400 errors etc.
You can also use the Moz Crawl Test (PRO only as well) which will return to you the HTTP status code of each URL crawled.
Optionally, you can also use Screaming Frog SEO Spider (free up to 500 pages to crawl, paid for bigger sites).
-Dan
-
Does anyone have any clue on how I can check my internal broken links using MOZ tools?
-
Hello,
If you have a full Sitemap, transform it to a .txt by putting only URL (Excel can do that easily) try Screaming Frog SEO on list mode, which is free regardless of the number of links.
Even if you search a free tool, I strongly recommand to invest 99£/year on this tool, it's worth it.
-
Ok, first, does the url with /errors/error_404 (without the query string) exist?
Also - are you using something to create your 404 pages for tracking purposes?
The query string on the url would appear to be stating the referring url anyway - hence ?q=-i3 basically equates to the page /i3
At a guess - it could be there is a script running to create logs on specific 404's, or create a new log each time one occurs - if this folder is visable to the crawlers, it would get spidered and subsequently the problem would arise.
-
Yeah I tried that but the URL in the URL coloumn shows for example:
/errors/error_404?q=-i3
and the Referrer shows:
/-i3
Neither page exists and so I do not know where the page which contains this broken link is...
-
It does, when you get the report, filter column D marked "4XX (Client Error)" and you will see your 404's there - further along in the report you will also be able to see referring url which will show you which page is linking to it.
-
Checked out the tool I just linked to on here and it doesn't include broken link data
-
Is the the custom crawl tool on here you are referring too? http://pro.seomoz.org/tools/crawl-test
-
Custom crawl right here on SEOmoz, or you could use Xenu
-
Hi Dan,
Thanks for that tool. It works great and especially like that it works in Chrome.
Do you however know of a tool which would do this site wide rather than per page?
-
This tool is a great extension for Chrome
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How can I stop a tracking link from being indexed while still passing link equity?
I have a marketing campaign landing page and it uses a tracking URL to track clicks. The tracking links look something like this: http://this-is-the-origin-url.com/clkn/http/destination-url.com/ The problem is that Google is indexing these links as pages in the SERPs. Of course when they get indexed and then clicked, they show a 400 error because the /clkn/ link doesn't represent an actual page with content on it. The tracking link is set up to instantly 301 redirect to http://destination-url.com. Right now my dev team has blocked these links from crawlers by adding Disallow: /clkn/ in the robots.txt file, however, this blocks the flow of link equity to the destination page. How can I stop these links from being indexed without blocking the flow of link equity to the destination URL?
Technical SEO | | UnbounceVan0 -
Abnormally high internal link reported in Google Search Console not matching Moz reports
If I'm looking at our internal link count and structure on Google Search Console, some pages are listed as having over a thousand internal links within our site. I've read that having too many internal links on a page devalues that page's PageRank, because the value is divided amongst the pages it links out to. Likewise, I've heard having too many internal links is just bad in general for SEO. Is that true? The problem I'm facing is determining how Google is "discovering" these internal links. If I'm just looking at one single page reported with, say, 1,350 links and I'm just looking at the code, it may only have 80 or 90 actual links. Moz will confirm this, as well. So why would Google Search Console report different? Should I be concerned about this?
Technical SEO | | Closetstogo0 -
How does link juice flow through hreflang?
We want to use the hreflang tag on our site (direct users searching for the Spanish version of spanishdict.com to spanishdict.com/traductor). Before doing so, we were wondering how link juice flows through hreflang? Any insight or resources on this would be very helpful. Thanks!
Technical SEO | | CuriosityMedia0 -
Links from Instructables.com?
This is a silly newbie question. But will posting on www.instructables.com with some valuable content and url link back to my site help with "linking"? Or do they put a no-follow on all links on their site? Thanks for answering! Ron
Technical SEO | | yatesandcojewelers0 -
What is link Schemes?
Hello Friends, Today I am reading about link schemes on http://support.google.com/webmasters/bin/answer.py?hl=en&answer=66356 there are a several ways how to avoid Google penalties and also talk about the low quality links. But I can't understand about "Low-quality directory or bookmark site links" Is there he talked about low page rank, Alexa or something else?
Technical SEO | | KLLC0 -
Link Volume - calculate what you need?
Hi everyone, an interesting question here. How do you determien what link volume you should try and get into your website? What analysis do you do to determine the number of links you feel is right to go into a back-link profiel every month? obviously there is no magic number but its an interesting question to know what others do. Obviously you don't want to build too many or too little. If you have been penalised for bad links in the past and are now back on track - how do you calculate the volume? Do you take links dropping out into consideration?
Technical SEO | | pauledwards0 -
Is link cloaking bad?
I have a couple of affiliate gaming sites and have been cloaking the links, the reason I do this is to stop have so many external links on my sites. In the robot.txt I tell the bots not to index my cloaked links. Is this bad, or doesnt it really matter? Thanks for your help.
Technical SEO | | jwdesign0 -
Add to Cart Link
We have shopping cart links (<a href's,="" not="" input="" buttons)="" that="" link="" to="" a="" url="" along="" the="" lines="" of="" cart="" add="" 123&return="/product/123. </p"></a> <a href's,="" not="" input="" buttons)="" that="" link="" to="" a="" url="" along="" the="" lines="" of="" cart="" add="" 123&return="/product/123. </p">The SEOMoz site crawls are flagging these as a massive number of 302 redirects and I also wonder what sort of effect this is having on linkjuice flowing around the site. </a> <a href's,="" not="" input="" buttons)="" that="" link="" to="" a="" url="" along="" the="" lines="" of="" cart="" add="" 123&return="/product/123. </p">I can see several possible solutions: Make the links nofollow Make the links input buttons Block /cart/add with robots.txt Make the links 301 instead of 302 Make the links javascript (probably worst care) All of these would result in an identical outcome for the UX, but are very different solutions. What would you suggest?</a>
Technical SEO | | Aspedia0