Rogerbot directives in robots.txt
-
I feel like I spend a lot of time setting false positives in my reports to ignore.
Can I prevent Rogerbot from crawling pages I don't care about with robots.txt directives? For example., I have some page types with meta noindex and it reports these to me. Theoretically, I can block Rogerbot from these with a robots,txt directive and not have to deal with false positives.
-
Yes, you can definitely use the robots.txt file to prevent Rogerbot from crawling pages that you don’t want to include in your reports. This approach can help you manage and minimize false positives effectively.
To block specific pages or directories from being crawled, you would add directives to your robots.txt file. For example, if you have certain page types that you’ve already set with meta noindex, you can specify rules like this:
User-agent: Rogerbot Disallow: /path-to-unwanted-page/ Disallow: /another-unwanted-directory/
This tells Rogerbot not to crawl the specified paths, which should reduce the number of irrelevant entries in your reports.
However, keep in mind that while robots.txt directives can prevent crawling, they do not guarantee that these pages won't show up in search results if they are linked from other sites or indexed by different bots.
Additionally, using meta noindex tags is still a good practice for pages that may occasionally be crawled but shouldn’t appear in search results. Combining both methods—robots.txt for crawling and noindex for indexing—provides a robust solution to manage your web presence more effectively.
-
Never mind, I found this. https://moz.com/help/moz-procedures/crawlers/rogerbot
-
@awilliams_kingston
Yes, you can use robots.txt directives to prevent Rogerbot from crawling certain pages or sections of your site, which can help reduce the number of false positives in your reports. By doing so, you can focus Rogerbot’s attention on the parts of your site that matter more to you and avoid reporting issues on pages you don't care about.Here’s a basic outline of how you can use robots.txt to block Rogerbot:
Locate or Create Your robots.txt File: This file should be placed in the root directory of your website (e.g., https://www.yourwebsite.com/robots.txt).
Add Directives to Block Rogerbot: You’ll need to specify the user-agent for Rogerbot and define which pages or directories to block. The User-agent directive specifies which web crawlers the rules apply to, and Disallow directives specify the URLs or directories to block.
Here’s an example of what your robots.txt file might look like if you want to block Rogerbot from crawling certain pages:
javascript
Disallow: /path-to-block/
Disallow: /another-path/
If you want to block Rogerbot from accessing pages with certain parameters or patterns, you can use wildcards:javascript
Disallow: /path-to-block/*
Disallow: /another-path/?parameter=
Verify the Changes: After updating the robots.txt file, you can use tools like Google Search Console or other site analysis tools to check if the directives are being applied as expected.Monitor and Adjust: Keep an eye on your reports and site performance to ensure that blocking these pages is achieving the desired effect without inadvertently blocking important pages.
By doing this, you should be able to reduce the number of irrelevant or false positive issues reported by Rogerbot and make your reporting more focused and useful.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Small startup Marketing leader - What are 3 actionable reports I can review daily
Hi all - I've joined a small startup as their first marketing hire and I am strategizing, planning, and executing all work. I need to get to 3-4 reports I focus on per channel so I can still be relatively effective across multiple channels. What are 3-4 reports I should be laser-focused on in Moz that will help me ID opportunities/threats and be able to identify best actions from.
Digital Marketing | | AndrewAeqium0 -
Client Dashboard Options
My company is an agency, and we manage many SEO Campaigns. I love the reports, but I'd really like to add an online dashboard that my clients can login to and see the same up to date stats as I do in Moz Pro. I can't do it with Seats because you can't limit a seat to a specific campaign (as far as I know). Has anyone found a solution for this?
Reporting & Analytics | | bizmarquee3 -
How do you report SEO audit findings?
Hello, Mozzers! I'm curious to know how you report SEO audit findings. Do you use a spreadsheet? A presentation? A formal report? Or maybe something else. If you have a favourite audit template, I'd love to see it. A second question: what things do you report in an audit? I currently report crawl findings, authority and trust, link profiles, and competitive analysis. I also investigate a site's security—that's not usually part of an audit, but site owners need to know about it. What do you report to your audit customers? Thanks for sharing your auditing wisdom!
Intermediate & Advanced SEO | | AndyKubrin0 -
Main Website Redirects to Mobile Website, Mobile Website counts this as direct traffic, is there a way to tell what the source/medium is?
Hello, The situation is that someone is arriving on my main website https://www.example.com and being redirected to http://m.example.com. When this happens my analytics says that the traffic is all direct coming to my mobile site. However, I know people clicking on my google cpc, and some google organic users are hitting the main website and being redirected. Before we didn't have as good of a redirect on our main website so I could tell organic and cpc traffic coming in, now my main website has a huge drop in these categories because they are redirecting to mobile but I can't tell on my mobile how much traffic from each is going to the mobile site. Is there a way to fix this? Is it because my main website is https:// and mobile is a http:// (as I know that sometimes makes traffic direct) or is it a bigger problem that can't be resolved? Thanks
Reporting & Analytics | | oxfordseminars0 -
Spike in Direct visits with Drop in Google Organic Visits
Does anyone have an explanation for why Google organic visits plummeted while direct visits rose the same amount. Total visits have been very normal. jI1Bn2B,UuzONNK kepV0eu
Reporting & Analytics | | phogan0 -
How to get a list of robots.txt file
This is my site. http://muslim-academy.com/ Its in wordpress.I just want to know is there any way I can get the list of blocked URL by Robots.txt In Google Webmaster its not showing up.Just giving the number of blocked URL's. Any plugin or Software to extract the list of blocked URL's.
Reporting & Analytics | | csfarnsworth0 -
Amazon.com inc.increase in direct traffic
Hi All, I have seen a increase of direct traffic from hostname amazon.com inc. This only happened on one day. Any ideas what/why it is? Thanks
Reporting & Analytics | | Sayers0 -
How much direct traffic is really direct?
Does anyone else think that a large chunk of traffic labelled as "Direct" in your analytics isn't direct at all. When you analyse traffic trends it seems that a large percentage could just be browsers with their referring URL hidden so it only appears direct. Here's the evidence: When we've been affected by major search algorithm changes, we've seen big changes in direct traffic as well as organic, but not in referral traffic. If direct traffic is just bookmarks, typed-in URLs, and people clicking through from emails why is direct traffic 85% new visitors? We don't do any offline advertising, so you'd expect genuine direct traffic to be returning visitors -- either our brand loyalists or subscribers to our email newsletters. If you segment direct traffic into new and returning visitors and look at a major algo update as discussed in 1), you find all the drop in direct traffic is from New Direct visitors, with no drop at all in Returning Direct visitors. Can anyone explain who these New, Direct visitors are if not simply mislabelled new, search visitors. Cookie deletion can't be the problem (ie: they can't be Returning, Direct really) because the traffic doesn't behave like returning, direct (that is, it varies too much). I'd be really interest to hear theories, and whether anyone has any figures on the extent of HTTP referrer blocking.
Reporting & Analytics | | Dennis-529610