SEO Best Practices regarding Robots.txt disallow
-
I cannot find hard and fast direction about the following issue:
It looks like the Robots.txt file on my server has been set up to disallow "account" and "search" pages within my site, so I am receiving warnings from the Google Search console that URLs are being blocked by Robots.txt. (Disallow: /Account/ and Disallow: /?search=). Do you recommend unblocking these URLs?
I'm getting a warning that over 18,000 Urls are blocked by robots.txt. ("Sitemap contains urls which are blocked by robots.txt"). Seems that I wouldn't want that many urls blocked. ?
Thank you!!
-
mmm it depends.
it's really hard for me to answer without knowing your site but I would say that you're in the good direction. You want to provide google more ways to reach your quality content.
Now do you have any other page that is bringing bots there via a normal user navigation or is it all search driven?
While google can crawl pages that discovered via internal/external links it can't reproduce searches by typing in your nav bar, so I doubt those pages should be extremely valuable unless you link to them somehow. In that case you may want to keep google crawling them.
A different thing would be if you want to "index" them, as being searches they are probably aggregating different information already present on the site. For indexation purposes you may want to keep them out of the index while still allowing the bot to run through them.
Again beware of the crawl budget, you don't want google to be wandering around millions of search results instead of your money pages, unless you're able to let them crawl only a sub portion of that.
I hope this made sense
-
Thank you for your response! I'm going to do a bit more research but I think I will disallow "account", but unblock "search". The search feature on my site pulls up quality content, so seems like I would want that to be crawled. Does this sound logical to you?
-
That could be completely normal. Google sends a warning because you're giving conflicting directions as you are preventing them to crawl pages (via robots) you asked them to index (via sitemap).
They do not know how important those pages may be for you so you are the one that needs to assess what to do net.
Are those pages important for you? Do you want them to be in the index? if that's the case change your robots.txt rule, if not then remove them from the sitemap.
About the previous answer robots text is not used to block hackers but quite the opposite. Hackers can easily find via the robots txt which are the pages you'd like to block and visit them as they may be key pages (ex. wp-admin), but let's not focus on that as hackers have so many ways to find core pages that it's not the topic. Robots txt is normally used to avoid duplication issues and to prevent google from crawling low value pages and waste crawl budget.
-
Typically, you only want robots.txt to block access points that would allow hackers into your site like an admin page (e.g. www.examplesite.com/admin/). You definitely don't want it blocking your whole site. A developer or webmaster would be better at speaking to the specifics, but that's the quick, high-level answer.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What is the best SEO tool to check internal linking structure
HI, Is there any tool to check how a website's internal linking structure has been linked. Some times few important pages may not linked very well and some links will be over linked. This will surge rankings...like if more links are pointing to one page? Is there any tool to check this?
Intermediate & Advanced SEO | | vtmoz0 -
SEO page descriptions on mobile - how to hide while preserving the juice for SEO?
Hi everybody, On our pages we have crafted good text paragraphs for SEO purposes. On desktop everything is fine but on mobile the paragraph of text pushes the main content really low on the page. Is there a way of hiding the text while preserving the SEO juices and not getting penalised by Google for spamming techniques? I'd appreciate any recommendations on how to deal with this. Thanks very much!
Intermediate & Advanced SEO | | Firebox0 -
Robots txt is case senstive? Pls suggest
Hi i have seen few urls in the html improvements duplicate titles Can i disable one of the below url in the robots.txt? /store/Solar-Home-UPS-1KV-System/75652
Intermediate & Advanced SEO | | Rahim119
/store/solar-home-ups-1kv-system/75652 if i disable this Disallow: /store/Solar-Home-UPS-1KV-System/75652 will the Search engines scan this /store/solar-home-ups-1kv-system/75652 im little confused with case senstive.. Pls suggest go ahead or not in the robots.txt0 -
Frames and SEO
In the old days frames were a problem because they could end up being orphaned pages that get indexed and appear in the SERPS but had to navigation so they were useless. Are frames still a problem with Google and if so what are the suggested solutions?
Intermediate & Advanced SEO | | siteoptimized0 -
SEO through Social Media, what have you been doing? :)
Hi, Just a general discussion really, what sort of thing have you been up to regarding social media and it helping SEO. One thing that does interest me is generating tweets/likes and also using Google plus profile to help SEO.
Intermediate & Advanced SEO | | activitysuper0 -
Local SEO Best Practices
Hello Everyone, I'm new to SEOmoz, I'm looking to use this as a tool to really help me, and evenually I can help others. I am an Web Developer with some online marketing experience. I did Local SEO a Few Years ago, and things have really changed since then. I know this Panda and Penguin update really is putting a hurting on the directory submission. Google no longer has 'Citations" on their places page, and many other changes. With that being said, what are some best practices for Local SEO? I am a propeller head by nature, but am also very creative when I need to be. I have potental sites to market, anywhere from Holistic Medical Doctors, Plastic Surgeons Community Blogs, and Auto Repair Shops, Law firms (to give you some perspectic) I also read Danny Dover's Book, to learn some more about SEO, the one thing that is unclear is how to acquire quality links I would really appreciate any perspective on this, every little thing helps Zach Russell
Intermediate & Advanced SEO | | ZacharyRussell0 -
Should we block urls like this - domainname/shop/leather-chairs.html?brand=244&cat=16&dir=ascℴ=price&price=1 within the robots.txt?
I've recently added a campaign within the SEOmoz interface and received an alarming number of errors ~9,000 on our eCommerce website. This site was built in Magento, and we are using search friendly url's however most of our errors were duplicate content / titles due to url's like: domainname/shop/leather-chairs.html?brand=244&cat=16&dir=asc&order=price&price=1 and domainname/shop/leather-chairs.html?brand=244&cat=16&dir=asc&order=price&price=4. Is this hurting us in the search engines? Is rogerbot too good? What can we do to cut off bots after the ".html?" ? Any help would be much appreciated 🙂
Intermediate & Advanced SEO | | MonsterWeb280