Get a list of robots.txt blocked URL and tell Google to crawl and index it.
-
Some of my key pages got blocked by robots.txt file and I have made required changes in robots.txt file but how can I get the blocked URL's list.
My webmaster page Health>blocked URL's shows only number not the blocked URL's.My first question is from where can I fetch these blocked URL's and how can I get them back in searches,
One other interesting point I see is that blocked pages are still showing up in searches.Title is appearing fine but Description shows blocked by robots.txt file.
I need urgent recommendation as I do not want to see drop in my traffic any more.
-
"changing the lastmod of those pages to today".
How can I make these changes?
Right now the news is that Resubmitted the Sitemap and no warnings this time.
-
I imagine that since you've got a robots txt error you'll probably ended closing a whole directory to bots which you wanted to be indexed. You can easily spot the directory and resubmit a sitemap to google changing the lastmod of those pages to today and the priority to 1 but only of those pages.
If you still receive warnings it may be due to errors in your sitemap. You're probably including some directory you don't want. You can try it in GWT putting in the box at the bottom the url you want to maintain in the index and then trying to see if some urls are being blocked by your robots.
If you want you can post here your robots and the URIs you want to be indexed without knowing the domain so that won't be public. Hope this may help you
-
Ok Resubmitted it.but even with updated file it gives a lot of errors.I think it takes some time.20,016 warnings
I have not added no index attribute in my header region.It was all messy stuff with robots.txt file.It means that with site showing up in SERP the rank will probably be the same or it was deranked?
-
Go into GWMT and resubmit sitemap.xml files (with the URLs you want indexed) for recrawling and Google will digest the sitemaps again, instead of waiting for Googlebot to come around on their own, you are requesting it to come around, also include those new sitemap files in your robots.txt file.
-
In Google Webmaster Tools, go to Health -> Fetch As Google. Then add the previously blocked URL and click Fetch. Once you've done that, refresh the page and click "Submit to index". That should get Google indexing those pages again.
Getting external links to your pages also helps get pages crawled & indexed, so it may be worth submitting your pages to social bookmarking sites, or get other types of backlinks to your previously blocked pages if possible.
-
Since you fixed your robots.txt file you should be good to. It will probably take a few days for Google to recrawl your site and update the index with the URLs they are now allow to crawl.
Blocked URLs can still show up in SERPs if you haven't defined the no-index attribute in your section.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Can I redo my submitted sitemap to Google?
We are a electronic hardware manufacture with a fairly large catalog of products. I dynamically built our site and we have over 705,000 unique products that we can offer. With our php framework I was able to create sitemaps that hold every product unique url. After doing all of that I submitted our data to Google. Then waited with a cocktail encouraged that we'd grow up the ranks of Google organically. Well, that didn't happen. Besides several other problems (lack of overall unique content, appearance of duplicate content, no meta description, no unique page titles, poor use of heading tags and no rel canonical tags) how can I get a "do-over" with Google and my submitted sitemaps? Can they be re-submitted? Can they even be deleted?
Reporting & Analytics | | jandk40140 -
Google Webmaster Tools During GA Transition?
I'm working with a client that is launching a new website. Google Webmaster Tools can just be disconnected, then reconnected to the new Google Analytics property, correct? Without any data loss in Webmaster Tools? Thanks! Becky
Reporting & Analytics | | Becky_Converge0 -
Save a Google Analytics account
We recently started with a client who has Google Analytics already installed on their current site, but they have no idea the account login for it? Is there anyway to transfer ownership of this just based on the code?
Reporting & Analytics | | WillWatrous0 -
Get anchor text, nofollow info etc from a list of links
Hi everybody. I'm currently doing a backlink audit for a client and I've hit a small problem. I'm combining data from Ahrefs, OSE, Webmaster Tools and Link Detox. I've got around 27k links in total now, but the issue is that WMT does not provide data on target page, anchor text and nofollow/dofollow. This means I have around 1k links with only partial information. Does anyone know of a way that I can get this data automatically? Thanks!
Reporting & Analytics | | Blink-SEO1 -
URL Structure for Study Guides
I'm creating study guides for various musicals, plays and operas. A few issues: I want to easily view in Google Analytics data for all study guides, data at the category level (musical,play, opera) and data for an individual show. I want urls to be as short as possible for usability purposes. Some show titles are the same for shows in different categories. For example, there is the play Romeo and Juliet but there is also an opera Romeo and Juliet. Each study guide contains multiple sections that will live on different URLs (overview, context, plot summary, characters) What would be the ideal URL structure? Here's what I was currently thinking we should use: /show/play/romeo-juliet/ /show/play/romeo-juliet/context /show/play/romeo-juliet/plot /show/play/romeo-juliet/characters
Reporting & Analytics | | stageagent0 -
Landing page URL appearing as keyword
Hi Mozers, I've recently experienced the URLs of my key landing pages coming up as keywords. This has been on the rise since early July (when it was relatively insignificant) to the current position (see image below) where they make up the majority of my top keywords. Drilling down into a bit more detail, this seems to be almost exclusively Desktop traffic but in terms of Technology there are no clear standouts (seems to be mostly Windows OS and Chrome). Has anyone else been experiencing this?
Reporting & Analytics | | mopland0 -
Google Webmaster not accounting for internal links
Hi SEO gurus! All my websites in GWT show the website in question at the top of the "Links to your site", in the form of: Domains Total links my-site.com 1,000 third-party-1.com 500 third-party-2.com 300 third-party-3.com 200 etc.com 100 However, I have a specific account that suddenly (a few weeks back) disappeared its own link count: Domains Total links third-party-1.com 500 third-party-2.com 300 third-party-3.com 200 etc.com 100 Has this happened to any of you? Any ideas how to solve it? The website is www.gmvbodybuilding.com which you can see has plenty of properly formed links.
Reporting & Analytics | | hectorpn0 -
Google analytics tracking
Is their a built-in or easy way to track the amount of times an exit link is clicked on a page and show that with other links of the page. For instance: I had a page with several external ad links on it and I want to track how many each got and rank them in google analytics
Reporting & Analytics | | insitegoogle0