Crawl Diagnostics Error Spike
-
With the last crawl update to one of my sites there was a huge spike in errors reported. The errors jumped by 16,659 -- majority of which are under the duplicate title and duplicate content category.
When I look at the specific issues it seems that the crawler is crawling a ton of blank pages on the sites blog through pagination.
The odd thing is that the site has not been updated in a while and prior to this crawl on Jun 4th there were no reports of these blank pages.
Is this something that can be an error on the crawler side of things?
Any suggestions on next steps would be greatly appreciated. I'm adding an image of the error spike
-
This would be another issue. I would need to look at the code to give you more insight. But off the bat I assume that this is an issue regarding mislabeling the rel=next and rel=prev. They can be kind of tricky to work with on a broad based update due to the fact that they are intended to refer to specific pages. If you do not have the end page labeled Google says :
"When implemented incorrectly, such as omitting an expected rel="prev" or rel="next" designation in the series, we'll continue to index the page(s), and rely on our own heuristics to understand your content."
I would look into this first. If the answer is still elusive to you the next option would probably be finding a different set of eyes on the code to see if there are any minor oversights that you may have overlooked.
-
One last thing;
It seems that I have a game plan for addressing this issue, but as I think about this one thing has me concerned in the way Roger crawled the site.
The site has maybe a total of 100 articles, which would account for ?Page=10, but what I'm seeing is errors on ?Page=104. When you look at that page its a blank. Where is Roger coming up with that parameter?
Do you think this is a Roger issue or something else?
-
Makes sense
-
Unless you have some super secret page that is buried somewhere deep down in your site that you can ONLY get to from those pages, it wouldn't make sense to have them follow the links. All that will happen is they land on the next page, scrape it to the noindex tag and move on. They won't index and this just waste your sites bandwidth and slows everything else down. If it's a noindex it should usually be a nofollow unless you are looking to track conversions or some other specific only navigable through those pages.
-
Hey Jake;
Whats your option of using "nofollow" vs "follow" on the pages i'm blocking from indexing? Is there a reason to prevent them from following the links on these pages?
-
Cool glad we could help!
if you want to clean up your code and are posting site wide for them I would recommend the none tag
Accounts for both
noindex, nofollow
-
Thank you again for the input, the goal here is not provide accurate reporting and ensure that the site conforms to the search engines requirements.
Currently the "?page=" parameter is not blocked through . it sounds like this maybe the issue.
I will update the code to address that and see what kind of results we get with the next update. I think this is best addressed at the code level, rather then the robots.txt.
Thanks
-
Rodger crawls like the Google bot and takes his hints from the robot.txt file. So whatever Rodger is seeing is usually what the other spiders are seeing as well. From time to time I have encountered slight glitches to the SEOmoz crawler as they change and update their algorithm.
When it comes down to it, Google examines a link profile through a microscope akin to the Large Hadron Collider. where as we have to examine it through a magnifying glass from 1935.
The wonderful people here at SEOmoz are always trying to give us a better view, but it is still imperfect. I would say if all else fails and this report continues to show errors in moz then get your reports for your clients directly from webmaster tools.
-
** How do I tell Roger no to crawl these blank pages?**
Any easy solution is to block roger in robots.txt
User-agent: rogerbot
Disallow: [enter pages you do not wish to be crawled]
But a better solution would be to fix the root problem. If your only goal is to provide clean reporting to your client the above will work. If your goal is to ensure your site is crawled correctly by Google/Bing, then Jake's suggestion will work. You can help Google and Bing understand your site by telling them how to handle parameters.
I would prefer to fix the root issue though. Do the pages which are being reported as duplicate content have the "noindex" tag on them? If so, you can report the issue to the moz help desk (help@seomoz.org) so they can investigate the problem.
-
Hey Jake;
Thanks for your feedback, i did make some changes to the code (posted in the reply to Jamie). I'll take a closer look at the webmaster tools to make sure things are OK on that end.
FYI: The "rel=prev / rel=next tags" are implemented
I added code to manage
to pages that are accessed through
- /Blog/?tag=
- /Blog/category/
- /Blog/archive.aspx
As a secondary concern, with Roger now reporting all these issues in SEOMoz, I provide these reports to my clients and thus having 16k errors is not a good PR thing. How do I tell Roger no to crawl these blank pages?
-
It looks like Rodger found his way into your variable URLs!
This could definitely cause a problem if the engine crawlers are seeing this path as well. Have you made any changes to the code on your site or the URL structure lately?
Regardless, you might want to examine in your Webmaster Tools for both Google and Bing.
For Google you will want to check the blocked URL's under the Health menu. This will give you the information on what pages are and are not blocked. If you notice that the Head Match term you are looking to exclude is not listed make sure that you upload the term to the robots.txt file on your site. Other fixes for this include canonicalisation tagging or the implementation of the rel=prev / rel=next tags. There are a few other ways that are more complicated and I recommend avoiding unless absolutely necessary.
But good news everyone! Google has a few ways to go about fixing the indexation.
Bing is a little Different but just as easy. In the Bing Webmaster Tools under the Index tab, there is a tool called URL Nor<a class="cpad Subject message-low-priority-icon marginleft5 bold">malization</a> you can tell the crawlers to exclude a portion of the query string without changing anything on your database. It also automatically finds and suggests <a class="cpad Subject message-low-priority-icon marginleft5 bold">query parameters for normalization as well. This is a recent change for Bing and could account for the sudden jump in warnings.</a>
I hope this helps and you keep being awesome!
-
Hey Jamie;
In an effort to block crawling of pages on the blog that are essentially duplicating content I added coded (on (4/16) to insert :
to pages that are accessed through
/Blog/?tag=
/Blog/category/
/Blog/archive.aspx
I did not do this for
/Blog/?page=
There were no changes to the robots.txt
There were no updates to canonical tag
There were no updates to pagination
Thanks for your prompt reply
-
Can you share what changes have been made to the site? A few ways this can happen are:
-
a change to the robots.txt file
-
a change to your site's template either removing a canonical tag, a noindex tag, or altering your pagination in any way such as modifying paginated titles
-
resolving an onsite issue which prevented crawling of these pages
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Htaccess and robots.txt and 902 error
Hi this is my first question in here I truly hope someone will be able to help. It's quite a detailed problem and I'd love to be able to fix it through your kind help. It regards htaccess files and robot.txt files and 902 errors. In October I created a WordPress website from what was previously a non-WordPress site it was quite dated. I had built the new site on a sub-domain I created on the existing site so that the live site could remain live whilst I created on the subdomain. The site I built on the subdomain is now live but I am concerned about the existence of the old htaccess files and robots txt files and wonder if I should just delete the old ones to leave the just the new on the new site. I created new htaccess and robots.txt files on the new site and have left the old htaccess files there. Just to mention that all the old content files are still sat on the server under a folder called 'old files' so I am assuming that these aren't affecting matters. I access the htaccess and robots.txt files by clicking on 'public html' via ftp I did a Moz crawl and was astonished to 902 network error saying that it wasn't possible to crawl the site, but then I was alerted by Moz later on to say that the report was ready..I see 641 crawl errors ( 449 medium priority | 192 high priority | Zero low priority ). Please see attached image. Each of the errors seems to have status code 200; this seems to be applying to mainly the images on each of the pages: eg domain.com/imagename . The new website is built around the 907 Theme which has some page sections on the home page, and parallax sections on the home page and throughout the site. To my knowledge the content and the images on the pages are not duplicated because I have made each page as unique and original as possible. The report says 190 pages have been duplicated so I have no clue how this can be or how to approach fixing this. Since October when the new site was launched, approx 50% of incoming traffic has dropped off at the home page and that is still the case, but the site still continues to get new traffic according to Google Analytics statistics. However Bing Yahoo and Google show a low level of Indexing and exposure which may be indicative of the search engines having difficulty crawling the site. In Google Analytics in Webmaster Tools, the screen text reports no crawl errors. W3TC is a WordPress caching plugin which I installed just a few days ago to speed up page speed, so I am not querying anything here about W3TC unless someone spots that this might be a problem, but like I said there have been problems re traffic dropping off when visitors arrive on the home page. The Yoast SEO plugin is being used. I have included information about the htaccess and robots.txt files below. The pages on the subdomain are pointing to the live domain as has been explained to me by the person who did the site migration. I'd like the site to be free from pages and files that shouldn't be there and I feel that the site needs a clean up as well as knowing if the robots.txt and htaccess files that are included in the old site should actually be there or if they should be deleted... ok here goes with the information in the files. Site 1) refers to the current website. Site 2) refers to the subdomain. Site 3 refers to the folder that contains all the old files from the old non-WordPress file structure. **************** 1) htaccess on the current site: ********************* BEGIN W3TC Browser Cache <ifmodule mod_deflate.c=""><ifmodule mod_headers.c="">Header append Vary User-Agent env=!dont-vary</ifmodule>
Moz Pro | | SEOguy1
<ifmodule mod_filter.c="">AddOutputFilterByType DEFLATE text/css text/x-component application/x-javascript application/javascript text/javascript text/x-js text/html text/richtext image/svg+xml text/plain text/xsd text/xsl text/xml image/x-icon application/json
<ifmodule mod_mime.c=""># DEFLATE by extension
AddOutputFilter DEFLATE js css htm html xml</ifmodule></ifmodule></ifmodule> END W3TC Browser Cache BEGIN W3TC CDN <filesmatch ".(ttf|ttc|otf|eot|woff|font.css)$"=""><ifmodule mod_headers.c="">Header set Access-Control-Allow-Origin "*"</ifmodule></filesmatch> END W3TC CDN BEGIN W3TC Page Cache core <ifmodule mod_rewrite.c="">RewriteEngine On
RewriteBase /
RewriteCond %{HTTP:Accept-Encoding} gzip
RewriteRule .* - [E=W3TC_ENC:_gzip]
RewriteCond %{HTTP_COOKIE} w3tc_preview [NC]
RewriteRule .* - [E=W3TC_PREVIEW:_preview]
RewriteCond %{REQUEST_METHOD} !=POST
RewriteCond %{QUERY_STRING} =""
RewriteCond %{REQUEST_URI} /$
RewriteCond %{HTTP_COOKIE} !(comment_author|wp-postpass|w3tc_logged_out|wordpress_logged_in|wptouch_switch_toggle) [NC]
RewriteCond "%{DOCUMENT_ROOT}/wp-content/cache/page_enhanced/%{HTTP_HOST}/%{REQUEST_URI}/_index%{ENV:W3TC_PREVIEW}.html%{ENV:W3TC_ENC}" -f
RewriteRule .* "/wp-content/cache/page_enhanced/%{HTTP_HOST}/%{REQUEST_URI}/_index%{ENV:W3TC_PREVIEW}.html%{ENV:W3TC_ENC}" [L]</ifmodule> END W3TC Page Cache core BEGIN WordPress <ifmodule mod_rewrite.c="">RewriteEngine On
RewriteBase /
RewriteRule ^index.php$ - [L]
RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule . /index.php [L]</ifmodule> END WordPress ....(((I have 7 301 redirects in place for old page url's to link to new page url's))).... #Force non-www:
RewriteEngine on
RewriteCond %{HTTP_HOST} ^www.domain.co.uk [NC]
RewriteRule ^(.*)$ http://domain.co.uk/$1 [L,R=301] **************** 1) robots.txt on the current site: ********************* User-agent: *
Disallow:
Sitemap: http://domain.co.uk/sitemap_index.xml **************** 2) htaccess in the subdomain folder: ********************* Switch rewrite engine off in case this was installed under HostPay. RewriteEngine Off SetEnv DEFAULT_PHP_VERSION 53 DirectoryIndex index.cgi index.php BEGIN WordPress <ifmodule mod_rewrite.c="">RewriteEngine On
RewriteBase /WPnewsiteDee/
RewriteRule ^index.php$ - [L]
RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule . /subdomain/index.php [L]</ifmodule> END WordPress **************** 2) robots.txt in the subdomain folder: ********************* this robots.txt file is empty **************** 3) htaccess in the Old Site folder: ********************* Deny from all *************** 3) robots.txt in the Old Site folder: ********************* User-agent: *
Disallow: / I have tried to be thorough so please excuse the length of my message here. I really hope one of you great people in the Moz community can help me with a solution. I have SEO knowledge I love SEO but I have not come across this before and I really don't know where to start with this one. Best Regards to you all and thank you for reading this. moz-site-crawl-report-image_zpsirfaelgm.jpg0 -
Since July 1, we've had a HUGE jump in errors on our weekly crawl. We don't think anything has changed on our website. Has MOZ changed something that would account for a large leap in duplicate content and duplicate title errors?
Our error report went from 1,900 to 18,000 in one swoop, starting right around the first of July. The errors are duplicate content and duplicate title, as if it does not see our 301 redirects. Any insights?
Moz Pro | | KristyFord0 -
Does Open Site Explorer purposefully not crawl some sites?
I use both SEOmoz's Open Site Explorer and Web Master Tools to find backlinks when conducting link audits. WMT always finds more links than OSE; I understand Google's database is bigger. But what is interesting to me is that it seems that a large percentage of the links WMT finds that OSE does not are real crappy links that I don't want. That makes me wonder if SEOmoz decides not to crawl certain, low quality, sites? Just curious.
Moz Pro | | ILM_Marketing0 -
Does crawling help in optimisation.?
the website is as it was last week. no optimisation from my side for 10 days now. i was ranked 5 with my keyword not much competition there. however 2 days ago i registrred at seomoz and created a campaign for my website with my keywords that were ranked 5 in search. today i see that my rank has gone up to 2. i have nt done any optimisation neither have ii created any backlinks. so how and why did i climb up? i just created a campaign and let seomoz crawl my website for 2days. am i to assume seomoz crawl optimises website? if that is the case then can i create a campaign crawl pages, climb up in searches, delete the campaign after a week, create it again crawl pages and climb up and so on ? please advise?
Moz Pro | | wahin10 -
How long does it take for a campaign website crawl to be completed?
Our campaign website crawl has been 'crawling' now for 5 days. Is this a normal phenomenon or is something hanging up?
Moz Pro | | Discountvc0 -
How come when I export a error list I can only export the first page?
I am working on fixing the 4xx errors. I have found the easiest way to do this would be to export the list, print it out, and check off the ones i've fixed. The site only lets me export the first page. We'll appreciate any help. Thanks, Ryan D. Gran --Not sure what category this question belongs in so selected SEOmoz Tools--
Moz Pro | | dggusmc0 -
Crawler reporting incorrect URLs, resulting in false errors...
The SEOmoz crawler is showing 236 Duplicate Page Titles. When I go in to see what page titles are duplicated I see that the URLs in question are incorrect and read "/about/about/..." instead of just "/about/" The shown page duplicates are the result of the crawler is ending up on the "Page not found" page. Could it be the result of using relative links on the site? Anything I can do to remedy? Thanks for your help! -Frank
Moz Pro | | Clements1 -
How long does a crawl take?
A crawl of my site started on the 8th July & is still going on - is there something wrong???
Moz Pro | | Brian_Worger1