OSE link report showing links to 404 pages on my site
-
I did a link analysis on this site mormonwiki.com. And many of the pages shown to be linked to were pages like these http://www.mormonwiki.com/wiki/index.php?title=Planning_a_trip_to_Rome_By_using_Movie_theatre_-_Your_five_Fun_Shows2052752
There happens to be thousands of them and these pages actually no longer exist but the links to them obviously still do. I am planning to proceed by disavowing these links to the pages that don't exist. Does anyone see any reason to not do this, or that doing this would be unnecessary?
Another issue is that Google is not really crawling this site, in WMT they are reporting to have not crawled a single URL on the site. Does anyone think the above issue would have something to do with this? And/or would you have any insight on how to remedy it?
-
The site does have and has had ranking issues since the first Penguin and has really had problems the last few months. And other than some minor things low quality links are really the only problem with the site.
-
Hi,
Adam is correct that the disavow tool should only be used if you think the links are causing you significant ranking problems. It's become quite common for people to disavow links without either a confirmed penalty or ranking issues, but those two factors were originally how Google recommended the tool be used.
What it sounds like has happened to your site with these bad pages is that spammers have created spam pages on the wiki then pointed links to those pages from elsewhere. It's a very common and old spam tactic, used on sites that allow UGC.
Those pages are now returning 404s, so technically the inbound links pointing to them should not hurt your website or cause a penalty. It's generally assumed the links to 404 pages (good or bad links) don't hurt or help. I disagree that they'll cause a "bad user experience" as it sounds like they have been built for spam purposes only - no one is going to try and visit these links.
If you believe these links are causing a ranking issue, the disavowal tool is certainly an option - I take it there's no chance you can negotiate these links' removal with the folks who built them? Removing links is always preferable to using disavowal also.
-
If you are seeing zero pages indexed and zero traffic from search then I would assume you have perhaps verified and subsequently are looking at data for the non-www version of the domain.
Double check that the site listed in WMT is www.mormonwiki.com and not mormonwiki.com. If you are looking at indexation and traffic data for the www version then there may be something else going on and unfortunately I wouldn't be able to diagnose the issue without looking at the WMT account.
Have your rankings been significantly affected? You would need to perform a fair amount of analysis before you can conclude that the site has been affected algorithmically. You would also need to be sure that any negative impact to rankings is a result of poor quality links and not something else, such as on-page factors.
Using the disavow should really be a last resort and only if it has been impossible to get troublesome links removed. As the warning from Google states, the disavow feature 'can potentially harm your site's performance' so I would not recommend using it until you have performed more in-depth analysis.
-
Right so if the pages no longer exist they need to be gotten rid of right? Most of these won't be removed by the webmasters and so they'll need to be disavowed right?
These pages were UGC and are essentially spam, and entirely irrelevant to anything on the site itself. So 301 redirects would not be wise or useful I don't think.
-
It hasn't received a manual action no. But that doesn't mean algorthimically the site isn't being affected.
So you're saying to not worry at all about these links?
They offer nothing in terms of value. If going to live pages they would be considered very spammy and completely irrelevant. But since these pages don't even exist you're saying it's unnecessary to bother with them at all?
I'm seeing the crawlability issue in WMT itself. The strange thing is that I know some pages have been indexed, we get most of our traffic organically from Google. But WMT shows zero pages indexed, zero traffic from search etc. The site has been verified as well.
-
I agree with Adam, if the links are natural then there is no need to disavow them.
However, if the links go to pages that no longer exist then it provides a poor user experience that can harm your rankings. Think of it like having dead links on your website. Have you set up 301 redirects for the pages that have become inactive? If not, set them up and make sure to redirect the pages to relevant areas of the website (no all to the homepage). Do this and the links should pass more juice and your website's performance should improve.
-
Are you performing a link analysis because the site received a manual action notification in WMT? If the site hasn't received a penalty then there is no need to use the disavow feature. As Google states:
'This is an advanced feature and should only be used with caution. If used incorrectly, this feature can potentially harm your site’s performance in Google’s search results. We recommend that you disavow backlinks only if you believe you have a considerable number of spammy, artificial, or low-quality links pointing to your site, and if you are confident that the links are causing issues for you. In most cases, Google can assess which links to trust without additional guidance, so most normal or typical sites will not need to use this tool.'
In terms of the crawlability of the site, where are you seeing WMT reporting to have not crawled a single page? A simple site: search of the mormonwiki.com domain returns about 65,600 results and I can't see any major issues that would prevent search engines from crawling the site. However, I would probably fix the issue with the robots.txt file. Currently, www.mormonwiki.com/robots.txt 301 redirects to www.mormonwiki.com/Robots.txt, which returns a 404 error.
Hope that helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Cached version of my site is not showing content?
Hi mozzers, I am a bit worried since I looked a cache version of my site and somehow content is partially showing up and navigation has completely disappeared. Where could this come from? What should I be doing? Thanks!
Intermediate & Advanced SEO | | Taysir0 -
E-Commerce Site Collection Pages Not Being Indexed
Hello Everyone, So this is not really my strong suit but I’m going to do my best to explain the full scope of the issue and really hope someone has any insight. We have an e-commerce client (can't really share the domain) that uses Shopify; they have a large number of products categorized by Collections. The issue is when we do a site:search of our Collection Pages (site:Domain.com/Collections/) they don’t seem to be indexed. Also, not sure if it’s relevant but we also recently did an over-hall of our design. Because we haven’t been able to identify the issue here’s everything we know/have done so far: Moz Crawl Check and the Collection Pages came up. Checked Organic Landing Page Analytics (source/medium: Google) and the pages are getting traffic. Submitted the pages to Google Search Console. The URLs are listed on the sitemap.xml but when we tried to submit the Collections sitemap.xml to Google Search Console 99 were submitted but nothing came back as being indexed (like our other pages and products). We tested the URL in GSC’s robots.txt tester and it came up as being “allowed” but just in case below is the language used in our robots:
Intermediate & Advanced SEO | | Ben-R
User-agent: *
Disallow: /admin
Disallow: /cart
Disallow: /orders
Disallow: /checkout
Disallow: /9545580/checkouts
Disallow: /carts
Disallow: /account
Disallow: /collections/+
Disallow: /collections/%2B
Disallow: /collections/%2b
Disallow: /blogs/+
Disallow: /blogs/%2B
Disallow: /blogs/%2b
Disallow: /design_theme_id
Disallow: /preview_theme_id
Disallow: /preview_script_id
Disallow: /apple-app-site-association
Sitemap: https://domain.com/sitemap.xml A Google Cache:Search currently shows a collections/all page we have up that lists all of our products. Please let us know if there’s any other details we could provide that might help. Any insight or suggestions would be very much appreciated. Looking forward to hearing all of your thoughts! Thank you in advance. Best,0 -
Pages that did NOT 301 redirect to the new site
Hi, Is there a tool out there that can tell me what pages did NOT 301 redirect to the new sites? I need something rather than going into google.com and typing in site:oldsite.com to see if it's still indexed and if it's not 301 redirecting.. I'm not sure if screaming frog can do that. Thanks.
Intermediate & Advanced SEO | | ggpaul5620 -
How can I fix "Too Many On Page Links"?
One of the warnings from SEO Moz says that we have "too many on page links" on a series of pages on my website. The pages it's giving me these warnings on are on my printing sample pages. I'm assuming that it's because of my left navigation. You can see an example here: http://www.3000doorhangers.com/door-hanger-design-samples/deck-and-fence-door-hanger-samples/ Any suggestions on how to fix this warning? Thanks!
Intermediate & Advanced SEO | | JimDirectMailCoach0 -
Why is Alt tag occurances so high in the on page reporter?
I am trying to understand why the on page reporter shows so many occurrences of alt tags containing my keywords. Can anyone shed any light? This is the URL that's in question. http://www.towelsrus.co.uk/towels-hand-towels/aztex/turkish-cotton-hand-towel_ct472bd182pd2745.htm yMlM5.png
Intermediate & Advanced SEO | | Towelsrus0 -
Handling Similar page content on directory site
Hi All, SEOMOZ is telling me I have a lot of duplicate content on my site. The pages are not duplicate, but very similar, because the site is a directory website with a page for cities in multiple states in the US. I do not want these pages being indexed and was wanting to know the best way to go about this. I was thinking I could do a rel ="nofollow" on all the links to those pages, but not sure if that is the correct way to do this. Since the folders are deep within the site and not under one main folder, it would mean I would have to do a disallow for many folders if I did this through Robots.txt. The other thing I am thinking of is doing a meta noindex, follow, but I would have to get my programmer to add a meta tag just for this section of the site. Any thoughts on the best way to achieve this so I can eliminate these dup pages from my SEO report and from the search engine index? Thanks!
Intermediate & Advanced SEO | | cchhita0 -
Should I build links to the home page or a url containing the keyword?
I run an IT company and the company name does not contain the key word I am trying to rank on. I also have a bunch of pages with page rank that containing the actual keywords, for example: http://www.mycompanyname.com/tech-support/locations/brighton My target keyword is "Tech Support Brighton" My Home page is PR4 and my location based pages are PR3. My plan was to build 3 or 4 location pages for the locations we provide tech support for and target location based keyword Anchor text to these URL's e.g "Tech Support Brighton" and then for the home page build links that have the anchor text "Tech Support". Does this sound sane? Many Thanks, K
Intermediate & Advanced SEO | | SEOKeith0 -
Link Juice - Lots of Pages
I have a site, PricesPrices.com where I'm steadily building inbound links and pagerank. I have about 4600 pages on the site, most of which are baby products in the baby gear sector. There are many outdated items that aren't really my focus, but do pop up in long-tail search queries from time to time. My question is a pretty basic one. Theoretically if a site has say 28/100 link juice, then as you go deeper and deeper into the site, the link juice is divided more and more. My question: Is this really true or just a concept? My thoughts are to hide many of the products that i don't really need to focus on therefor passing more link juice to the products that remain, but I also don't want to that if it won't necessarily make the remaining pages rank higher or have more link juice. I also have to keep in mind the merchandising aspect of the site and providing a good user experience. If i only have 300 products on the site, there will be a ton of unhappy people who can't find the products they are looking for. Any thoughts and/or pointers in the direction of funneling that pagerank down into my site would be much appreciated. Thanks!
Intermediate & Advanced SEO | | modparent0