612 : Page banned by error response for robots.txt
-
Hi all,
I ran a crawl on my site https://www.drbillsukala.com.au and received the following error "612 : Page banned by error response for robots.txt."Before anyone mentions it, yes, I have been through all the other threads but they did not help me resolve this issue.
I am able to view my robots.txt file in a browser https://www.drbillsukala.com.au/robots.txt.
The permissions are set to 644 on the robots.txt file so it should be accessible
My Google Search Console does not show any issues with my robots.txt file
I am running my site through StackPath CDN but I'm not inclined to think that's the culpritOne thing I did find odd is that even though I put in my website with https protocol (I double checked), on the Moz spreadsheet it listed my site with http protocol.
I'd welcome any feedback you might have. Thanks in advance for your help.
Kind regards -
Hey there! Tawny from Moz's Help Team here.
After doing some quick searching, it looks like how you configure the rules for WAFs depends on what service you're using to host those firewalls. You may need to speak to their support team to ask how to configure things to allow our user-agents.
Sorry I can't be more help here! If you still have questions we can help with, feel free to reach out to us at help@moz.com and we'll do our best to assist you.
-
Hi, I am having the same issue.
Can you please tell me how you have created rule in Web Application Firewall to allow user agents rogerbot and dotbot.
Thanks!!
-
Hi Federico,
Thanks for the prompt. Yes, this solution worked. I'm hopeful that this thread helps others too because when I was troubleshooting the problem, the other threads were not helpful for my particular situation.
Cheers
-
Hi, did the solution work?
-
Hi Federico,
I think I have found the solution for this problem and am hopeful the crawl will be successful this time around. Based on further digging and speaking to the team at StackPath CDN, I have done the following:
- I added the following to my robots.txt file
User-agent: rogerbot
Disallow:User-agent: dotbot
Disallow:- I added a custom robots.txt file in my CDN which includes the above and then created a rule in my Web Application Firewall which allows user agents rogerbot and dotbot.
I'll let you know if the crawl was successful or not.
Kind regards
-
Thanks for your response Federico. I have checked my robots.txt tester in my Google Search Console and it said "allowed."
Oddly, it also happened on another site of mine that I'm also running through StackPath CDN with a web application firewall in place. This makes me wonder if perhaps the CDN/WAF are the culprits (?).
I'll keep poking around to see what I find.
Cheers -
Seems like an issue with the Moz crawler, as the robots.txt has no issues and the site loads just fine.
If you already tested your robots.txt using the Google Webmaster Tools "robots.txt Tester" just to be sure, then you should contact Moz here: https://moz.com/help/contact/pro
Hope it helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Unsolved Why my website is giving 4xx error
I was analyzing the website link, my website is giving me 4xx error. Google search console in not giving such error 27b42282-a5f6-4ad0-956a-91838633d5ad-image.png Any suggestion will be helpful. The site is on wordpress
Link Explorer | | VS-Gary0 -
Page Authority higher on my old (redirected) domain than new domain?
Hi everyone. I moved the domain my blog was on about 18 months ago, and it's ranking in Google never recovered. I've noticed in the 'Inbound Links' tool that my old domain (jamescrowley.co.uk) shows a page authority of 26, while my new domain (jamescrowley.net) shows a page authority of 15. Any idea why that would be? 301 redirects have been in place for the whole 18 months, so I'd have thought the authority would have been 'passed on' by now? Many thanks James
Link Explorer | | james.crowley0 -
Crawl a node js page - Why can I only see my frontpage?
Hi When i am trying to crawl my website ( https://www.doorot.com/ ) it can only find my frontpage. It's a node js page. Any one had the same problem or know how to crawl my site in order to see all my pages? Kasper
Link Explorer | | KasperClio1 -
Moz Pro: Linking RDs to Page much lower than Google Search Console
I'm trying to use the Analyze Keyword tool in Moz Pro, and in the SERP Analysis table, my page has a PA of one, and zero root domains linking to it. If I look at the page in Google Search console, it says I have 229 root domains linking to the page from well known domains like github.com, meetup.com, stackoverflow.com, etc. This particular keyword has been tracked in Moz for the last 6 months, but I just noticed that it was extremely low. I am relatively new to Moz, so forgive me if I sound confused, but can someone explain to me how the numbers can be so low?
Link Explorer | | jakebellacera0 -
803 Errors, how to deal with this?
Hello, During my last two MOZ crawls over couple of hundred 803 errors showed up. Thought that this might stop, but nop still creeping up. Not sure what has caused this. My server providers are WP-Engine and they already said that all good at their end. Pretty much all of those errors are for photos on my blog. I'm a photographer. I have a web guy as well, but he is not sure what to do now and how to get this fixed. Website is a-fotografy.co.uk Thank you and if someone could shed some light. I did research here already, but nothing what cover photo side. Regards, Armands
Link Explorer | | A_Fotografy0 -
804 error preventing website being crawled
Hi For both subdomains https://us.sagepub.com and https://uk.sagepub.com crawling is being prevented by a 804 error. I can't see any reason why this should be so as all content is served through https. Thanks
Link Explorer | | philmoorse0 -
SEO report for specific pages (not just based on keyword analysis)
Hi, i am after a report where i can type in a URL nd get SPECIFIC details on how to exactly optimise a SPECIFIC page (not just the home page) for better organic results. A complete overview to everything in relation to tags, images tags, headers etc etc. I am yet to find the ability to get such a report with your service. COudl you please advise how i can do so. Thanks
Link Explorer | | andrewlos0 -
Open Site Explorer Top Pages https Title tag issue
Hello, everybody. I've noticed this strange thing in Top Pages reporting of Open Site Explorer. If i do a report for any website which doesn't have security certificate installed, such as http://www.hyperlinksmedia.com I get normal results, with title tag shown for every page, but, if i do a report for website with Security certificate, such as https://www.hyperlinksmedia.com, report is coming back for non-secure URL version (http:...), so, it says "No Title" for any pages. I wonder, if it influence PA and DA scores. Although, if i run full crawl test, then i shows all meta and title tags for SSL version of a website's url. Thanks!
Link Explorer | | seomozinator0