OSE link report showing links to 404 pages on my site
-
I did a link analysis on this site mormonwiki.com. And many of the pages shown to be linked to were pages like these http://www.mormonwiki.com/wiki/index.php?title=Planning_a_trip_to_Rome_By_using_Movie_theatre_-_Your_five_Fun_Shows2052752
There happens to be thousands of them and these pages actually no longer exist but the links to them obviously still do. I am planning to proceed by disavowing these links to the pages that don't exist. Does anyone see any reason to not do this, or that doing this would be unnecessary?
Another issue is that Google is not really crawling this site, in WMT they are reporting to have not crawled a single URL on the site. Does anyone think the above issue would have something to do with this? And/or would you have any insight on how to remedy it?
-
The site does have and has had ranking issues since the first Penguin and has really had problems the last few months. And other than some minor things low quality links are really the only problem with the site.
-
Hi,
Adam is correct that the disavow tool should only be used if you think the links are causing you significant ranking problems. It's become quite common for people to disavow links without either a confirmed penalty or ranking issues, but those two factors were originally how Google recommended the tool be used.
What it sounds like has happened to your site with these bad pages is that spammers have created spam pages on the wiki then pointed links to those pages from elsewhere. It's a very common and old spam tactic, used on sites that allow UGC.
Those pages are now returning 404s, so technically the inbound links pointing to them should not hurt your website or cause a penalty. It's generally assumed the links to 404 pages (good or bad links) don't hurt or help. I disagree that they'll cause a "bad user experience" as it sounds like they have been built for spam purposes only - no one is going to try and visit these links.
If you believe these links are causing a ranking issue, the disavowal tool is certainly an option - I take it there's no chance you can negotiate these links' removal with the folks who built them? Removing links is always preferable to using disavowal also.
-
If you are seeing zero pages indexed and zero traffic from search then I would assume you have perhaps verified and subsequently are looking at data for the non-www version of the domain.
Double check that the site listed in WMT is www.mormonwiki.com and not mormonwiki.com. If you are looking at indexation and traffic data for the www version then there may be something else going on and unfortunately I wouldn't be able to diagnose the issue without looking at the WMT account.
Have your rankings been significantly affected? You would need to perform a fair amount of analysis before you can conclude that the site has been affected algorithmically. You would also need to be sure that any negative impact to rankings is a result of poor quality links and not something else, such as on-page factors.
Using the disavow should really be a last resort and only if it has been impossible to get troublesome links removed. As the warning from Google states, the disavow feature 'can potentially harm your site's performance' so I would not recommend using it until you have performed more in-depth analysis.
-
Right so if the pages no longer exist they need to be gotten rid of right? Most of these won't be removed by the webmasters and so they'll need to be disavowed right?
These pages were UGC and are essentially spam, and entirely irrelevant to anything on the site itself. So 301 redirects would not be wise or useful I don't think.
-
It hasn't received a manual action no. But that doesn't mean algorthimically the site isn't being affected.
So you're saying to not worry at all about these links?
They offer nothing in terms of value. If going to live pages they would be considered very spammy and completely irrelevant. But since these pages don't even exist you're saying it's unnecessary to bother with them at all?
I'm seeing the crawlability issue in WMT itself. The strange thing is that I know some pages have been indexed, we get most of our traffic organically from Google. But WMT shows zero pages indexed, zero traffic from search etc. The site has been verified as well.
-
I agree with Adam, if the links are natural then there is no need to disavow them.
However, if the links go to pages that no longer exist then it provides a poor user experience that can harm your rankings. Think of it like having dead links on your website. Have you set up 301 redirects for the pages that have become inactive? If not, set them up and make sure to redirect the pages to relevant areas of the website (no all to the homepage). Do this and the links should pass more juice and your website's performance should improve.
-
Are you performing a link analysis because the site received a manual action notification in WMT? If the site hasn't received a penalty then there is no need to use the disavow feature. As Google states:
'This is an advanced feature and should only be used with caution. If used incorrectly, this feature can potentially harm your site’s performance in Google’s search results. We recommend that you disavow backlinks only if you believe you have a considerable number of spammy, artificial, or low-quality links pointing to your site, and if you are confident that the links are causing issues for you. In most cases, Google can assess which links to trust without additional guidance, so most normal or typical sites will not need to use this tool.'
In terms of the crawlability of the site, where are you seeing WMT reporting to have not crawled a single page? A simple site: search of the mormonwiki.com domain returns about 65,600 results and I can't see any major issues that would prevent search engines from crawling the site. However, I would probably fix the issue with the robots.txt file. Currently, www.mormonwiki.com/robots.txt 301 redirects to www.mormonwiki.com/Robots.txt, which returns a 404 error.
Hope that helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Site shows up after re-indexing, then disappears.
I have a site, natvest.com, with which I sell real estate in Alabama and Georgia. I need to show up in an "Alabama Land for Sale" search. Same thing for Georgia. If I re-index my site, I show up for roughly one day, before disappearing again. Happens every time I re-index. Ideas?
Intermediate & Advanced SEO | | natvest0 -
How do we decide which pages to index/de-index? Help for a 250k page site
At Siftery (siftery.com) we have about 250k pages, most of them reflected in our sitemap. Though after submitting a sitemap we started seeing an increase in the number of pages Google indexed, in the past few weeks progress has slowed to a crawl at about 80k pages, and in fact has been coming down very marginally. Due to the nature of the site, a lot of the pages on the site likely look very similar to search engines. We've also broken down our sitemap into an index, so we know that most of the indexation problems are coming from a particular type of page (company profiles). Given these facts below, what do you recommend we do? Should we de-index all of the pages that are not being picked up by the Google index (and are therefore likely seen as low quality)? There seems to be a school of thought that de-indexing "thin" pages improves the ranking potential of the indexed pages. We have plans for enriching and differentiating the pages that are being picked up as thin (Moz itself picks them up as 'duplicate' pages even though they're not. Thanks for sharing your thoughts and experiences!
Intermediate & Advanced SEO | | ggiaco-siftery0 -
Cache and index page of Mobile site
Hi, I want to check cache and index page of mobile site. I am checking it on mobile phone but it is showing the cache version of desktop. So anybody can tell me the way(tool, online tool etc.) to check mobile site index and cache page.
Intermediate & Advanced SEO | | vivekrathore0 -
Wordpress site, MOZ showing missing meta description but pages do not exist on backend
I've got a wordpress website (a client) and MOZ keeps showing missing meta descriptions. When I look at the pages these are nonsense pages, they do exist somewhere but I am not seeing them on the backend. Questions: 1) how do I fix this? Maybe it's a rel con issue? why is this referring to "non-sense" pages? When I go to the page there is nothing on it except maybe an image or the headline, it's very strange. Any input out there I greatly appreciate. Thank you
Intermediate & Advanced SEO | | SOM240 -
How to Destroy Old 404 Pages
Hello Mozzers, So I just purchased a new domain and to my surprise it has a domain authority of 13 right out of the box (what luck!). I needed to investigate. To make a long story short the domain used to be home to a music blog that had hundreds of pages which of course are all missing now. I have about 400 pages on my hands that are resulting in a 404. How or what is the best method for eliminating these pages. Does deleting the Crawl Errors in Google Webmaster Tools do anything? Thanks
Intermediate & Advanced SEO | | TheOceanAgency0 -
Linking to bad sites
Hi, I just have a quick question. Is it very negative to link to "bad" sites, such as online pharmacies, dating, adult sites, that sort of stuff? How much does linking to a "bad" site negatively affect a "good" site? Thank you.
Intermediate & Advanced SEO | | salvyy0 -
Google consolidating link juice on duplicate content pages
I've observed some strange findings on a website I am diagnosing and it has led me to a possible theory that seems to fly in the face of a lot of thinking: My theory is:
Intermediate & Advanced SEO | | James77
When google see's several duplicate content pages on a website, and decides to just show one version of the page, it at the same time agrigates the link juice pointing to all the duplicate pages, and ranks the 1 duplicate content page it decides to show as if all the link juice pointing to the duplicate versions were pointing to the 1 version. EG
Link X -> Duplicate Page A
Link Y -> Duplicate Page B Google decides Duplicate Page A is the one that is most important and applies the following formula to decide its rank. Link X + Link Y (Minus some dampening factor) -> Page A I came up with the idea after I seem to have reverse engineered this - IE the website I was trying to sort out for a client had this duplicate content, issue, so we decided to put unique content on Page A and Page B (not just one page like this but many). Bizarrely after about a week, all the Page A's dropped in rankings - indicating a possibility that the old link consolidation, may have been re-correctly associated with the two pages, so now Page A would only be getting Link Value X. Has anyone got any test/analysis to support or refute this??0 -
Should I build links to the home page or a url containing the keyword?
I run an IT company and the company name does not contain the key word I am trying to rank on. I also have a bunch of pages with page rank that containing the actual keywords, for example: http://www.mycompanyname.com/tech-support/locations/brighton My target keyword is "Tech Support Brighton" My Home page is PR4 and my location based pages are PR3. My plan was to build 3 or 4 location pages for the locations we provide tech support for and target location based keyword Anchor text to these URL's e.g "Tech Support Brighton" and then for the home page build links that have the anchor text "Tech Support". Does this sound sane? Many Thanks, K
Intermediate & Advanced SEO | | SEOKeith0