OSE link report showing links to 404 pages on my site
-
I did a link analysis on this site mormonwiki.com. And many of the pages shown to be linked to were pages like these http://www.mormonwiki.com/wiki/index.php?title=Planning_a_trip_to_Rome_By_using_Movie_theatre_-_Your_five_Fun_Shows2052752
There happens to be thousands of them and these pages actually no longer exist but the links to them obviously still do. I am planning to proceed by disavowing these links to the pages that don't exist. Does anyone see any reason to not do this, or that doing this would be unnecessary?
Another issue is that Google is not really crawling this site, in WMT they are reporting to have not crawled a single URL on the site. Does anyone think the above issue would have something to do with this? And/or would you have any insight on how to remedy it?
-
The site does have and has had ranking issues since the first Penguin and has really had problems the last few months. And other than some minor things low quality links are really the only problem with the site.
-
Hi,
Adam is correct that the disavow tool should only be used if you think the links are causing you significant ranking problems. It's become quite common for people to disavow links without either a confirmed penalty or ranking issues, but those two factors were originally how Google recommended the tool be used.
What it sounds like has happened to your site with these bad pages is that spammers have created spam pages on the wiki then pointed links to those pages from elsewhere. It's a very common and old spam tactic, used on sites that allow UGC.
Those pages are now returning 404s, so technically the inbound links pointing to them should not hurt your website or cause a penalty. It's generally assumed the links to 404 pages (good or bad links) don't hurt or help. I disagree that they'll cause a "bad user experience" as it sounds like they have been built for spam purposes only - no one is going to try and visit these links.
If you believe these links are causing a ranking issue, the disavowal tool is certainly an option - I take it there's no chance you can negotiate these links' removal with the folks who built them? Removing links is always preferable to using disavowal also.
-
If you are seeing zero pages indexed and zero traffic from search then I would assume you have perhaps verified and subsequently are looking at data for the non-www version of the domain.
Double check that the site listed in WMT is www.mormonwiki.com and not mormonwiki.com. If you are looking at indexation and traffic data for the www version then there may be something else going on and unfortunately I wouldn't be able to diagnose the issue without looking at the WMT account.
Have your rankings been significantly affected? You would need to perform a fair amount of analysis before you can conclude that the site has been affected algorithmically. You would also need to be sure that any negative impact to rankings is a result of poor quality links and not something else, such as on-page factors.
Using the disavow should really be a last resort and only if it has been impossible to get troublesome links removed. As the warning from Google states, the disavow feature 'can potentially harm your site's performance' so I would not recommend using it until you have performed more in-depth analysis.
-
Right so if the pages no longer exist they need to be gotten rid of right? Most of these won't be removed by the webmasters and so they'll need to be disavowed right?
These pages were UGC and are essentially spam, and entirely irrelevant to anything on the site itself. So 301 redirects would not be wise or useful I don't think.
-
It hasn't received a manual action no. But that doesn't mean algorthimically the site isn't being affected.
So you're saying to not worry at all about these links?
They offer nothing in terms of value. If going to live pages they would be considered very spammy and completely irrelevant. But since these pages don't even exist you're saying it's unnecessary to bother with them at all?
I'm seeing the crawlability issue in WMT itself. The strange thing is that I know some pages have been indexed, we get most of our traffic organically from Google. But WMT shows zero pages indexed, zero traffic from search etc. The site has been verified as well.
-
I agree with Adam, if the links are natural then there is no need to disavow them.
However, if the links go to pages that no longer exist then it provides a poor user experience that can harm your rankings. Think of it like having dead links on your website. Have you set up 301 redirects for the pages that have become inactive? If not, set them up and make sure to redirect the pages to relevant areas of the website (no all to the homepage). Do this and the links should pass more juice and your website's performance should improve.
-
Are you performing a link analysis because the site received a manual action notification in WMT? If the site hasn't received a penalty then there is no need to use the disavow feature. As Google states:
'This is an advanced feature and should only be used with caution. If used incorrectly, this feature can potentially harm your site’s performance in Google’s search results. We recommend that you disavow backlinks only if you believe you have a considerable number of spammy, artificial, or low-quality links pointing to your site, and if you are confident that the links are causing issues for you. In most cases, Google can assess which links to trust without additional guidance, so most normal or typical sites will not need to use this tool.'
In terms of the crawlability of the site, where are you seeing WMT reporting to have not crawled a single page? A simple site: search of the mormonwiki.com domain returns about 65,600 results and I can't see any major issues that would prevent search engines from crawling the site. However, I would probably fix the issue with the robots.txt file. Currently, www.mormonwiki.com/robots.txt 301 redirects to www.mormonwiki.com/Robots.txt, which returns a 404 error.
Hope that helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do we decide which pages to index/de-index? Help for a 250k page site
At Siftery (siftery.com) we have about 250k pages, most of them reflected in our sitemap. Though after submitting a sitemap we started seeing an increase in the number of pages Google indexed, in the past few weeks progress has slowed to a crawl at about 80k pages, and in fact has been coming down very marginally. Due to the nature of the site, a lot of the pages on the site likely look very similar to search engines. We've also broken down our sitemap into an index, so we know that most of the indexation problems are coming from a particular type of page (company profiles). Given these facts below, what do you recommend we do? Should we de-index all of the pages that are not being picked up by the Google index (and are therefore likely seen as low quality)? There seems to be a school of thought that de-indexing "thin" pages improves the ranking potential of the indexed pages. We have plans for enriching and differentiating the pages that are being picked up as thin (Moz itself picks them up as 'duplicate' pages even though they're not. Thanks for sharing your thoughts and experiences!
Intermediate & Advanced SEO | | ggiaco-siftery0 -
Do I have to optimize every page on my site?
Hi guys I run my own photography webstie (www.hemeravisuals.co.uk Going through the process optimizing my page for seo. I have one question I have a few gallery pages with no text etc? Do I still have to optimize these ? Would it rank my site lower if they weren't optimized? And how can i do this sucessfully with little text on these pages ( I have indepth text on these subjects on my services & pricing pages? Kind Regards Cam
Intermediate & Advanced SEO | | hemeravisuals0 -
PDF Cached by Google, but not showing as link
The following pdf is cached by google: http://www.sba.gov/sites/default/files/files/REFERRAL%20LIST%20OF%20BOND%20AGENCIES_Florida.pdf However, OpenSiteExplorer is not listing any of the links as found in it. With such an authoritative site, I would think Google would value this, right? None of the sites listed rank well though and OpenSiteExplorer's inability to see the links makes me wonder if Google provides these sites any value at all. Is there any link juice or brand mention value here for Google?
Intermediate & Advanced SEO | | TheDude0 -
Best possible linking on site with 100K indexed pages
Hello All, First of all I would like to thank everybody here for sharing such great knowledge with such amazing and heartfelt passion.It really is good to see. Thank you. My story / question: I recently sold a site with more than 100k pages indexed in Google. I was allowed to keep links on the site.These links being actual anchor text links on both the home page as well on the 100k news articles. On top of that, my site syndicates its rss feed (Just links and titles, no content) to this page. However, the new owner made a mess, and now the site could possibly be seen as bad linking to my site. Google tells me within webmasters that this particular site gives me more than 400K backlinks. I have NEVER received one single notice from Google that I have bad links. That first. But, I was worried that this page could have been the reason why MY site tanked as bad as it did. It's the only source linking so massive to me. Just a few days ago, I got in contact with the new site owner. And he has taken my offer to help him 'better' his site. Although getting the site up to date for him is my main purpose, since I am there, I will also put effort in to optimizing the links back to my site. My question: What would be the best to do for my 'most SEO gain' out of this? The site is a news paper type of site, catering for news within the exact niche my site is trying to rank. Difference being, his is a news site, mine is not. It is commercial. Once I fix his site, there will be regular news updates all within the niche we both are in. Regularly as in several times per day. It's news. In the niche. Should I leave my rss feed in the side bars of all the content? Should I leave an achor text link on the sidebar (on all news etc.) If so: there can be just one keyword... 407K pages linking with just 1 kw?? Should I keep it to just one link on the home page? I would love to hear what you guys think. (My domain is from 2001. Like a quality wine. However, still tanked like a submarine.) ALL SEO reports I got here are now Grade A. The site is finally fully optimized. Truly nice to have that confirmation. Now I hope someone will be able to tell me what is best to do, in order to get the most SEO gain out of this for my site. Thank you.
Intermediate & Advanced SEO | | richardo24hr0 -
Google places page related places links to competitor
I noticed on a lot of Google places pages i create for my clients Google seems to put related places links at the bottom of the page which links directly to their competitors. how can i remove control or avoid these links been placed? Also any tips on improving the places page would be greatly appreciated thanks
Intermediate & Advanced SEO | | Bristolweb0 -
Do I need a canonical tag on the 404 error page?
Per definition, a 404 is displayed for different url (any not existing url ...). As I try to clean my website following SEOmoz pro advices, SEOmoz notify me of duplicate content on urls leading to a 404 🙂 This is I guess not that important, but just curious: should we add a cononical tag to the template returning the 404, with a canonical url such as www.mysite.com/404 ?
Intermediate & Advanced SEO | | nuxeo0 -
Can obfuscated Javascript be used for too many links on a page?
Hi mozzers Just looking for opinions/answers on if it is ever appropriate to use obfuscated Javascript on links when a page has many links but they need to be there for usability? It seems grey/black hat to me as it shows users something different to Google (alarm bells are sounding already!) BUT if the page has many links it's losing juice which could be saved....... Any thoughts appreciated, thanks.
Intermediate & Advanced SEO | | TrevorJones0 -
Should I build links to the home page or a url containing the keyword?
I run an IT company and the company name does not contain the key word I am trying to rank on. I also have a bunch of pages with page rank that containing the actual keywords, for example: http://www.mycompanyname.com/tech-support/locations/brighton My target keyword is "Tech Support Brighton" My Home page is PR4 and my location based pages are PR3. My plan was to build 3 or 4 location pages for the locations we provide tech support for and target location based keyword Anchor text to these URL's e.g "Tech Support Brighton" and then for the home page build links that have the anchor text "Tech Support". Does this sound sane? Many Thanks, K
Intermediate & Advanced SEO | | SEOKeith0