Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Should I disavow links from pages that don't exist any more
-
Hi. Im doing a backlinks audit to two sites, one with 48k and the other with 2M backlinks. Both are very old sites and both have tons of backlinks from old pages and websites that don't exist any more, but these backlinks still exist in the Majestic Historic index. I cleaned up the obvious useless links and passed the rest through Screaming Frog to check if those old pages/sites even exist.
There are tons of link sending pages that return a 0, 301, 302, 307, 404 etc errors. Should I consider all of these pages as being bad backlinks and add them to the disavow file?
Just a clarification, Im not talking about l301-ing a backlink to a new target page. Im talking about the origin page generating an error at ping eg: originpage.com/page-gone sends me a link to mysite.com/product1. Screamingfrog pings originpage.com/page-gone, and returns a Status error. Do I add the originpage.com/page-gone in the disavow file or not?
Hope Im making sense
-
Sounds a plan. Thanks for your help bud, much appreciated.
-
My take, I'll just go ahead and start doing other things to improve it's current rankings. I could assign someone to go over links if another team member is available.
If I see improvements, within the next month, then that's a good sign already that you should continue and not worry about the dead links.
It takes google a long time to actually forget about those links pointing to your site. So if they are dead AND then you didnt notice any increases or drops in analytics, then they are pretty much ineffective so they shouldnt be a major obstacle. I think someone coined a term for it, ghost links or something. LOL.
-
Hi. I did go through GA several years back, think back to 2011, but didn't really see dramatic changes in traffic other than a general trend of just low organic traffic throughout. Keep in mind that it's an engineering site, so no thousands of visit per day... the keywords that are important for the site get below 1000 searcher per month (data from the days when Google Keyword Tool shared this info with us mortals).
That said, I do notice in roughly 60% of the links absolutely no regard for anchors, so some are www.domain.com/index.php, Company Name, some are Visit Site, some are Website etc. Some anchors are entire generic sentences like "your company provided great service, your entire team should be commended blah blah blah". And there are tons of backlinks from http://jennifers.tempdomainname.com...a domain that a weird animal as there's not much data on who they are, what they do and what the deal is with the domain name itself. Weird.
In all honesty, nothing in WMT or GA suggests that the site got hit by either Penguin or Panda....BUT, having a ton of links that originate from non-existing pages, pages with no thematic proximity to the client site, anchors that are as generic as "Great Service"...is it a plus to err on the side of caution and get them disavowed, or wait for a reason from Google and then do the link hygiene?
-
Hi Igor,
Seeing ezinearticles in there is definitely a red flag that tells you that it probably has web directories, article networks, blog networks, pliggs, guestbooks and other links from that time.
Maybe you can dig up some old analytics data, check out when the traffic dropped.
If you did not see any heavy anchor text usage, then the site must've gotten away with a sitewide penalty, I would assume it's just a few (or many, but not all) of the keywords that got hit so either way, youll need to clean up -> disavow the links if they are indeed like that. So that's probably a reason for it's low organic rankings.
That, and since it's old, it might have been affected by panda too.
-
Thanks for your response. Im about done with cleaning up the link list in very broad strokes, eliminating obvious poor quality links, so in a few hours I could have a big list for disavowing.
The site is very specific, mechanical engineering thing and they sell technology and consulting to GM, GE, Intel, Nasa... so backlinks from sites for rental properties and resorts do look shady....even if they do return a 200 status.
But...how vigilent is google now with all the Penguin updates about backlinks from non-related sites, and my client's site has tons of them? And if Majestic reports them to have zero trust flow, is there a benefit of having them at all?
Thanks.
-
Hi. Thanks for responding. WMT shows just a fraction of the links actually. about few thousand for the site that Majestic Historic reports 48k. But I dont have any notifications of issues. Im guessing that with all the Penguin updates most sites won't get any notifications and it's up to us SEO guys to figure out why rankings are so low.
About quality of the links, many do come from weird sites, and I've noticed ezinearticles too. Problem is that the 48k portfolio was built by non-seo experts and now, few years after the fact, Im stuck with a site that doesn't rank well and has no notifications in WMT. But can I take the lack of notification as evidence that the site has no backlinks problem, or do I read-in the problem in poor organic ranking?
-
If I would be in that similar situation I would not really care about it but if it didn’t took too much of my time, I would have included all of these in the disavow file too.
But if the page is not giving a 200 status, this shouldn’t really be a problem.
Hope this helps!
-
Hi Igor,
Do they still show up in Webmaster tools? Do you have a penalty because of those links that used to link to the site? If not then I wouldn't really worry about it and just prioritize other things and make that a side task.
Are the majority of them on bad looking domains? If you checked the link URL on archive.org, were they spammy links? Then go ahead and include them in the disavow list.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Why can't google mobile friendly test access my website?
getting the following error when trying to use google mobile friendly tool: "page cannot be reached. This could be because the page is unavailable or blocked by robots.txt" I don't have anything blocked by robots.txt or robots tag. i also manage to render my pages on google search console's fetch and render....so what can be the reason that the tool can't access my website? Also...the mobile usability report on the search console works but reports very little, and the google speed test also doesnt work... Any ideas to what is the reason and how to fix this? LEARN MOREDetailsUser agentGooglebot smartphone
Technical SEO | | Nadav_W0 -
Blog Page Titles - Page 1, Page 2 etc.
Hi All, I have a couple of crawl errors coming up in MOZ that I am trying to fix. They are duplicate page title issues with my blog area. For example we have a URL of www.ourwebsite.com/blog/page/1 and as we have quite a few blog posts they get put onto another page, example www.ourwebsite.com/blog/page/2 both of these urls have the same heading, title, meta description etc. I was just wondering if this was an actual SEO problem or not and if there is a way to fix it. I am using Wordpress for reference but I can't see anywhere to access the settings of these pages. Thanks
Technical SEO | | O2C0 -
Should I noindex my blog's tag, category, and author pages
Hi there, Is it a good idea to no index tag, category, and author pages on blogs? The tag pages sometimes have duplicate content. And the category and author pages aren't really optimized for any search term. Just curious what others think. Thanks!
Technical SEO | | Rignite0 -
Why are these internal pages not showing any internal links?
If you look at Author profile pages like this one, http://experts.allbusiness.com/author/denise-oberry (THE top contributor on the site with over 82 posts under her belt), or any Author profile page, they show zero internal links or Page Authority. The same goes for most posts for each author on the site. Author pages should show internal links from every post the author has on the site. And specific posts should also have internal links from categories, etc. Yet they show zero. The only posts that show internal links and PA are ones that were either syndicated to the root domain's homepage, or syndicated to Fox Small Business. ZERO internal links. Does anyone know why this is? The root domain does not act this way with Author pages and posts. And I see nothing blocking links or indexing via the robots.txt file or page level nofollow tags. A real head scratcher for this SEO nerd, that I'm sure someone here will have a really simple answer to.
Technical SEO | | MiguelSalcido0 -
How Does Google's "index" find the location of pages in the "page directory" to return?
This is my understanding of how Google's search works, and I am unsure about one thing in specific: Google continuously crawls websites and stores each page it finds (let's call it "page directory") Google's "page directory" is a cache so it isn't the "live" version of the page Google has separate storage called "the index" which contains all the keywords searched. These keywords in "the index" point to the pages in the "page directory" that contain the same keywords. When someone searches a keyword, that keyword is accessed in the "index" and returns all relevant pages in the "page directory" These returned pages are given ranks based on the algorithm The one part I'm unsure of is how Google's "index" knows the location of relevant pages in the "page directory". The keyword entries in the "index" point to the "page directory" somehow. I'm thinking each page has a url in the "page directory", and the entries in the "index" contain these urls. Since Google's "page directory" is a cache, would the urls be the same as the live website (and would the keywords in the "index" point to these urls)? For example if webpage is found at wwww.website.com/page1, would the "page directory" store this page under that url in Google's cache? The reason I want to discuss this is to know the effects of changing a pages url by understanding how the search process works better.
Technical SEO | | reidsteven750 -
Google insists robots.txt is blocking... but it isn't.
I recently launched a new website. During development, I'd enabled the option in WordPress to prevent search engines from indexing the site. When the site went public (over 24 hours ago), I cleared that option. At that point, I added a specific robots.txt file that only disallowed a couple directories of files. You can view the robots.txt at http://photogeardeals.com/robots.txt Google (via Webmaster tools) is insisting that my robots.txt file contains a "Disallow: /" on line 2 and that it's preventing Google from indexing the site and preventing me from submitting a sitemap. These errors are showing both in the sitemap section of Webmaster tools as well as the Blocked URLs section. Bing's webmaster tools are able to read the site and sitemap just fine. Any idea why Google insists I'm disallowing everything even after telling it to re-fetch?
Technical SEO | | ahockley0 -
Adding 'NoIndex Meta' to Prestashop Module & Search pages.
Hi Looking for a fix for the PrestaShop platform Look for the definitive answer on how to best stop the indexing of PrestaShop modules such as "send to a friend", "Best Sellers" and site search pages. We want to be able to add a meta noindex ()to pages ending in: /search?tag=ball&p=15 or /modules/sendtoafriend/sendtoafriend-form.php We already have in the robot text: Disallow: /search.php
Technical SEO | | reallyitsme
Disallow: /modules/ (Google seems to ignore these) But as a further tool we would like to incude the noindex to all these pages too to stop duplicated pages. I assume this needs to be in either the head.tpl or the .php file of each PrestaShop module.? Or is there a general site wide code fix to put in the metadata to apply' Noindex Meta' to certain files. Current meta code here: Please reply with where to add code and what the code should be. Thanks in advance.0 -
Do or don't —forward a parked domain to a live website?
Hi all, I'm new to SEO and excited to see the launch of this forum. I've searched for an answer to this question but haven't been able to find out. I "attended" two webinars recently regarding SEO. The above subject was raised in each one and the speakers gave a polar opposite recommendations. So I'm completely at a loss as to what to do with some domains that are related to a domain used on a live website that I'm working to improve the SEO on. The scenario: Live website at (fictitious) www.digital-slr-camera-company.com. I also have 2 related domain names which are parked with the registrar: www.dslr.com, www.digitalslr.com. The question: Is there any SEO benefit to be gained by pointing the two parked domains to the website at www.digitalcamercompany.com? If so, what method of "pointing" should be used? Thanks to any and all input.
Technical SEO | | Technical_Contact0