Get a list of robots.txt blocked URL and tell Google to crawl and index it.
-
Some of my key pages got blocked by robots.txt file and I have made required changes in robots.txt file but how can I get the blocked URL's list.
My webmaster page Health>blocked URL's shows only number not the blocked URL's.My first question is from where can I fetch these blocked URL's and how can I get them back in searches,
One other interesting point I see is that blocked pages are still showing up in searches.Title is appearing fine but Description shows blocked by robots.txt file.
I need urgent recommendation as I do not want to see drop in my traffic any more.
-
"changing the lastmod of those pages to today".
How can I make these changes?
Right now the news is that Resubmitted the Sitemap and no warnings this time.
-
I imagine that since you've got a robots txt error you'll probably ended closing a whole directory to bots which you wanted to be indexed. You can easily spot the directory and resubmit a sitemap to google changing the lastmod of those pages to today and the priority to 1 but only of those pages.
If you still receive warnings it may be due to errors in your sitemap. You're probably including some directory you don't want. You can try it in GWT putting in the box at the bottom the url you want to maintain in the index and then trying to see if some urls are being blocked by your robots.
If you want you can post here your robots and the URIs you want to be indexed without knowing the domain so that won't be public. Hope this may help you
-
Ok Resubmitted it.but even with updated file it gives a lot of errors.I think it takes some time.20,016 warnings
I have not added no index attribute in my header region.It was all messy stuff with robots.txt file.It means that with site showing up in SERP the rank will probably be the same or it was deranked?
-
Go into GWMT and resubmit sitemap.xml files (with the URLs you want indexed) for recrawling and Google will digest the sitemaps again, instead of waiting for Googlebot to come around on their own, you are requesting it to come around, also include those new sitemap files in your robots.txt file.
-
In Google Webmaster Tools, go to Health -> Fetch As Google. Then add the previously blocked URL and click Fetch. Once you've done that, refresh the page and click "Submit to index". That should get Google indexing those pages again.
Getting external links to your pages also helps get pages crawled & indexed, so it may be worth submitting your pages to social bookmarking sites, or get other types of backlinks to your previously blocked pages if possible.
-
Since you fixed your robots.txt file you should be good to. It will probably take a few days for Google to recrawl your site and update the index with the URLs they are now allow to crawl.
Blocked URLs can still show up in SERPs if you haven't defined the no-index attribute in your section.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Strange landing page in Google Analytics
Hello MOZ Community, The website in question is https://x-y.com/ When i looked at the landing pages report in GA , x-y.com is appended at the end of every URL like this. https://x-y.com/x-y.com When i open the above URL in GA interface, it shows page not found. This is obvious as there is no such URL.
Reporting & Analytics | | Johnroger
The metrics like sessions, Users, Bounce rate all look good. In the property settings, The default URL is written like this http:// cell-gate.com (Please note that s is missing in property settings). But how is traffic tracked correctly How do i solve this problem. What settings should we change to make the landing pages report look ok Thanks0 -
Google Search Console - Click Count Inconsistency
In Google's search console I see a discrepancy between click counts. At top I see this here and then beneath I see these kinds of numbers for click counts here. So the top click count says 252 and the bottom section appears to only shows less than 40. Probably a simple explanation here that I'm just not seeing. Thanks!
Reporting & Analytics | | a_toohill0 -
I get a - 'Temporarily unreachable' error message when I 'Fetch as Google' Any ideas please??
I wanted to Fetch this page and got this error from Google - Temporarily unreachable. I've never had this issue before?? I checked another page and it came back as 'Complete', so no problems there? Any ideas? Thank you in advance.
Reporting & Analytics | | MissThumann0 -
Google Tag Manager Tracking
We have Google Tag Manager set up to track our website https://kontenta.co.uk/ Although we have a blog - https://kontenta.co.uk/designroom/ I'm just wondering if it's better to track the sites separately? If so, how do I add the blog to tag manager? Does it need setting up as a filter to match the regular expression https://kontenta.co.uk/designroom/* We just need to figure out the best way to measure both the website and blog traffic and if users are coming from the blog to the site, and what interactions they are making. Thanks
Reporting & Analytics | | Jseddon920 -
Time to get reconsideratio n request?
Hi there, recently I have been getting failed reconsideration requests back within 7 days. The last one i submitted (third) is now 2 weeks and no message back. Does this mean do you think it's gone to the next stage of auditing? is it a good sign? Typically when failed I have always has that response within 7 days. I also noticed a slight improvement in rankings but no message back in webmaster tools. Does anyone have experience with this? if so does this sound about right? thanks guys.
Reporting & Analytics | | pauledwards0 -
How to hook up a ppc campaign to a google + Page
Greetings,
Reporting & Analytics | | Nightwing
Sometimes you just want to give Google a big slap for making straight forward requests damn impossible. So all i ma trying to ad is point a ppc ad at this Google + account <a>https://plus.google.com/118393512656496298734#118393512656496298734/posts</a> But i get a warning sign saying:
"The URL must be for a Google+ page, not a personal profile" I then spend half an hour tring to find a Google + page but get no where fast 😞 Warning message illustrated here:
http://i216.photobucket.com/albums/cc53/zymurgy_bucket/google-page-plus_zps46ff995a.jpg So my question is please how to a get the Google + page for this account:
<a>https://plus.google.com/118393512656496298734#118393512656496298734/posts</a> Any insights welcome!
David0 -
Search Engine blocked by robots.txt
I am getting this error whe I try to crawl http://photosales.belfasttelegraph.co.uk/ but my robots.txt file does not block any bots?
Reporting & Analytics | | MirandaP0 -
Why do I get lots of traffic from a bizarre keyword?
Bit of an odd one but I've been getting a large and steady stream of traffic over the last few months from a very random keyword that according to addwords figures shows "on data". Its our second biggest referring term only beaten by our brand name. We get more traffic from this term than keywords we have invested a lot of time in that show thousands of traffic volume in addwords. When looking at behavioral data its gets odder, a bounce rate of 98.11% time on site 2 seconds and page visits 1.02. So this traffic isn't real traffic and it's not real people. So my questions are, what is it? why do we get this random traffic, has anyone els noticed things like this and is it a problem? I presume it must be something to do with some sort of spam but apart from that i'm stumped. It's just one of those things that has been bugging me so I would appreciate any help. Kind Regards Paul
Reporting & Analytics | | pauldoffman0