Find broken links in Excel?
-
Hello,
I have a large list of URL's in an excel sheet and I am looking for a way to check them for 404 errors. Please help!
- Adam
-
If this is something you check for often (which it probably should be), you might want to check out a program like Screaming Frog. That makes it really easy to find 404s and any other response codes.
-
If you want to use Excel, check out this macro which should be what you're looking for - http://stackoverflow.com/questions/1118221/sort-dead-hyperlinks-in-excel-with-vba
-
Do a Google Search for "bulk url checker" and a ton of results come up (most you can export back to excel).
Also, possible solution here: http://www.mrexcel.com/forum/excel-questions/507688-excel-url-check-help.html
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
how to find broken outbound links ?
hey guy, can anyone help me in finding broken outbound link on my website by using moz ? does Moz has this function ?
Technical SEO | | rogerdam0 -
Regarding Internal Links
I analyse my Birthday Page "https://www.giftalove.com/birthday"with comapare link profiles and found that total Internal Link 47,234. How my internal link suddenly increse. Please provide my details about my internal links.
Technical SEO | | Packersmove0 -
Link Spam from Competitor Help
A clients link profile is recently getting lots of spam links related to "abortion pills" and "does my husband cheat" I found a few of the sites that link, and it appears that there is some malicious code on the site injecting links at the top of the site. http://www.med-reporter.at/index.asp?men=Gesundheit&submen=Produkte&artid=1587&kategorie=&blockzl=3 Can anyone look at the link above and tell me what network or software is creating these links?
Technical SEO | | webbroi0 -
Reciprocal links and nofollow tag
What happens if I link to a site using the nofollow tag and they are linking back to me with a dofollow link? Will it give me as much power and link juice as if it was a one way link (to me) or will Google discount the link because it's reciprocal?
Technical SEO | | Livet0 -
No crawl code for pages of helpful links vs. no follow code on each link?
Our college website has many "owners" who want pages of "helpful links" resulting in a large number of outbound links. If we add code to the pages to prevent them from being crawled, will that be just as effective as making every individual link no follow?
Technical SEO | | LAJN0 -
Crawl Diagnostics - How to find where broken links are located?
Hi, One of my sites has a 4xx error that has been picked up in the crawl diagnostics section. It is a broken link. Does anybody know if it is possible for me to find out which page the broken link was found on? I have checked all of the pages on the site that I thought were linking to the page that seems to have a problem but all of these links are fine / not broken. Any ideas? Thanks
Technical SEO | | CherryK0 -
301s and Link Juice
So I know that a 301 will pass the majority of link juice to the new site, but if that 301 is taken away what happens?
Technical SEO | | kylesuss0 -
External Links from own domain
Hi all, I have a very weird question about external links to our site from our own domain. According to GWMT we have 603,404,378 links from our own domain to our domain (see screen 1) We noticed when we drilled down that this is from disabled sub-domains like m.jump.co.za. In the past we used to redirect all traffic from sub-domains to our primary www domain. But it seems that for some time in the past that google had access to crawl some of our sub-domains, but in december 2010 we fixed this so that all sub-domain traffic redirects (301) to our primary domain. Example http://m.jump.co.za/search/ipod/ redirected to http://www.jump.co.za/search/ipod/ The weird part is that the number of external links kept on growing and is now sitting on a massive number. On 8 April 2011 we took a different approach and we created a landing page for m.jump.co.za and all other requests generated 404 errors. We added all the directories to the robots.txt and we also manually removed all the directories from GWMT. Now 3 weeks later, and the number of external links just keeps on growing: Here is some stats: 11-Apr-11 - 543 747 534 12-Apr-11 - 554 066 716 13-Apr-11 - 554 066 716 14-Apr-11 - 554 066 716 15-Apr-11 - 521 528 014 16-Apr-11 - 515 098 895 17-Apr-11 - 515 098 895 18-Apr-11 - 515 098 895 19-Apr-11 - 520 404 181 20-Apr-11 - 520 404 181 21-Apr-11 - 520 404 181 26-Apr-11 - 520 404 181 27-Apr-11 - 520 404 181 28-Apr-11 - 603 404 378 I am now thinking of cleaning the robots.txt and re-including all the excluded directories from GWMT and to see if google will be able to get rid of all these links. What do you think is the best solution to get rid of all these invalid pages. moz1.PNG moz2.PNG moz3.PNG
Technical SEO | | JacoRoux0