The Moz Q&A Forum

    • Forum
    • Questions
    • Users
    • Ask the Community

    Welcome to the Q&A Forum

    Browse the forum for helpful insights and fresh discussions about all things SEO.

    1. SEO and Digital Marketing Forum
    2. Categories
    3. SEO Tactics
    4. Reporting & Analytics
    5. Rogerbot directives in robots.txt

    Moz Q&A is closed.

    After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.

    Rogerbot directives in robots.txt

    Reporting & Analytics
    rogerbotreporting
    4 3 374
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as question
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • awilliams_kingston
      awilliams_kingston last edited by

      I feel like I spend a lot of time setting false positives in my reports to ignore.

      Can I prevent Rogerbot from crawling pages I don't care about with robots.txt directives? For example., I have some page types with meta noindex and it reports these to me. Theoretically, I can block Rogerbot from these with a robots,txt directive and not have to deal with false positives.

      khaansaab21 awilliams_kingston MarkWade 3 Replies Last reply Reply Quote 0
      • MarkWade
        MarkWade @awilliams_kingston last edited by

        @awilliams_kingston

        Yes, you can definitely use the robots.txt file to prevent Rogerbot from crawling pages that you don’t want to include in your reports. This approach can help you manage and minimize false positives effectively.

        To block specific pages or directories from being crawled, you would add directives to your robots.txt file. For example, if you have certain page types that you’ve already set with meta noindex, you can specify rules like this:

        User-agent: Rogerbot
        Disallow: /path-to-unwanted-page/
        Disallow: /another-unwanted-directory/
        

        This tells Rogerbot not to crawl the specified paths, which should reduce the number of irrelevant entries in your reports.

        However, keep in mind that while robots.txt directives can prevent crawling, they do not guarantee that these pages won't show up in search results if they are linked from other sites or indexed by different bots.

        Additionally, using meta noindex tags is still a good practice for pages that may occasionally be crawled but shouldn’t appear in search results. Combining both methods—robots.txt for crawling and noindex for indexing—provides a robust solution to manage your web presence more effectively.

        1 Reply Last reply Reply Quote 0
        • awilliams_kingston
          awilliams_kingston @awilliams_kingston last edited by

          Never mind, I found this. https://moz.com/help/moz-procedures/crawlers/rogerbot

          1 Reply Last reply Reply Quote 0
          • khaansaab21
            khaansaab21 @awilliams_kingston last edited by

            @awilliams_kingston
            Yes, you can use robots.txt directives to prevent Rogerbot from crawling certain pages or sections of your site, which can help reduce the number of false positives in your reports. By doing so, you can focus Rogerbot’s attention on the parts of your site that matter more to you and avoid reporting issues on pages you don't care about.

            Here’s a basic outline of how you can use robots.txt to block Rogerbot:

            Locate or Create Your robots.txt File: This file should be placed in the root directory of your website (e.g., https://www.yourwebsite.com/robots.txt).

            Add Directives to Block Rogerbot: You’ll need to specify the user-agent for Rogerbot and define which pages or directories to block. The User-agent directive specifies which web crawlers the rules apply to, and Disallow directives specify the URLs or directories to block.

            Here’s an example of what your robots.txt file might look like if you want to block Rogerbot from crawling certain pages:

            javascript

            Disallow: /path-to-block/
            Disallow: /another-path/
            If you want to block Rogerbot from accessing pages with certain parameters or patterns, you can use wildcards:

            javascript

            Disallow: /path-to-block/*
            Disallow: /another-path/?parameter=
            Verify the Changes: After updating the robots.txt file, you can use tools like Google Search Console or other site analysis tools to check if the directives are being applied as expected.

            Monitor and Adjust: Keep an eye on your reports and site performance to ensure that blocking these pages is achieving the desired effect without inadvertently blocking important pages.

            By doing this, you should be able to reduce the number of irrelevant or false positive issues reported by Rogerbot and make your reporting more focused and useful.

            1 Reply Last reply Reply Quote 0
            • 1 / 1
            • First post
              Last post

            Got a burning SEO question?

            Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.


            Start my free trial


            Explore more categories

            • Moz Tools

              Chat with the community about the Moz tools.

              Getting Started
              Moz Pro
              Moz Local
              Moz Bar
              API
              What's New

            • SEO Tactics

              Discuss the SEO process with fellow marketers

              Content Development
              Competitive Research
              Keyword Research
              Link Building
              On-Page Optimization
              Technical SEO
              Reporting & Analytics
              Intermediate & Advanced SEO
              Image & Video Optimization
              International SEO
              Local SEO

            • Community

              Discuss industry events, jobs, and news!

              Moz Blog
              Moz News
              Industry News
              Jobs and Opportunities
              SEO Learn Center
              Whiteboard Friday

            • Digital Marketing

              Chat about tactics outside of SEO

              Affiliate Marketing
              Branding
              Conversion Rate Optimization
              Web Design
              Paid Search Marketing
              Social Media

            • Research & Trends

              Dive into research and trends in the search industry.

              SERP Trends
              Search Behavior
              Algorithm Updates
              White Hat / Black Hat SEO
              Other SEO Tools

            • Support

              Connect on product support and feature requests.

              Product Support
              Feature Requests
              Participate in User Research

            • See all categories

            • The Moz.com bot is overloading my server
              paulavervo
              paulavervo
              0
              8
              873

            • Zero '0' Total Visits
              rh-digi
              rh-digi
              0
              2
              1.3k

            • 520 Error from crawl report with Cloudflare
              awilliams_kingston
              awilliams_kingston
              0
              7
              2.0k

            • Landing pages report has no data even if I have ranking keywords and traffic
              davidevans_seo
              davidevans_seo
              0
              20
              4.5k

            • Can you accurately track lead source when using a cookie consent system?
              Marketing-EdJones
              Marketing-EdJones
              0
              3
              3.4k

            • Abnormally High Direct Traffic Volume
              graceflack 0
              graceflack 0
              1
              3
              3.6k

            • How to get a list of robots.txt file
              csfarnsworth
              csfarnsworth
              0
              6
              966

            Get started with Moz Pro!

            Unlock the power of advanced SEO tools and data-driven insights.

            Start my free trial
            Products
            • Moz Pro
            • Moz Local
            • Moz API
            • Moz Data
            • STAT
            • Product Updates
            Moz Solutions
            • SMB Solutions
            • Agency Solutions
            • Enterprise Solutions
            • Digital Marketers
            Free SEO Tools
            • Domain Authority Checker
            • Link Explorer
            • Keyword Explorer
            • Competitive Research
            • Brand Authority Checker
            • Local Citation Checker
            • MozBar Extension
            • MozCast
            Resources
            • Blog
            • SEO Learning Center
            • Help Hub
            • Beginner's Guide to SEO
            • How-to Guides
            • Moz Academy
            • API Docs
            About Moz
            • About
            • Team
            • Careers
            • Contact
            Why Moz
            • Case Studies
            • Testimonials
            Get Involved
            • Become an Affiliate
            • MozCon
            • Webinars
            • Practical Marketer Series
            • MozPod
            Connect with us

            Contact the Help team

            Join our newsletter
            Moz logo
            © 2021 - 2026 SEOMoz, Inc., a Ziff Davis company. All rights reserved. Moz is a registered trademark of SEOMoz, Inc.
            • Accessibility
            • Terms of Use
            • Privacy

            Looks like your connection to Moz was lost, please wait while we try to reconnect.