Staging & Development areas should be not indexable (i.e. no followed/no index in meta robots etc)
-
Hi
I take it if theres a staging or development area on a subdomain for a site, who's content is hence usually duplicate then this should not be indexable i.e. (no-indexed & nofollowed in metarobots) ? In order to prevent dupe content probs as well as non project related people seeing work in progress or finding accidentally in search engine listings ?
Also if theres no such info in meta robots is there any other way it may have been made non-indexable, or at least dupe content prob removed by canonicalising the page to the equivalent page on the live site ?
In the case in question i am finding it listed in serps when i search for the staging/dev area url, so i presume this needs urgent attention ?
Cheers
Dan
-
- use robots.txt vs the meta tags - robots.txt is preferred.
-
I'm about to issue these instructions would appreciate it if you could quickly confirm covers your advice correctly and nothing missing:
1) Setup a completely different GWT account unrelated to the main site, so that there is a new GWT account specific to the staging subdomain
2) Add a robots.txt on the staging area subdomain site that disallows all pages and all crawlers OR use the noindex meta tag on all pages. Its obviously very important when you update the main site it DOES NOTinclude or push out these files too (since that would result in main site or pages being de-indexed)3) Request removal of all pages in GWT. Leave the form blank for the page to be removed since this will remove the entire site4) After about 1 month (or you see that the pages are all out of the serps), and google has spidered and seen the robots.txt, then put up a password on the entire staging site.Note:For brand new sites staging areas that don't yet exist or exist but are new and not yet showing up in the index then simply add a password for human access to prevent the above process being required in the future. -
Thanks for clarifying that CleverPHD & thanks again for all your help and great advice
Have a great weekend !!
All Best
Dan
-
That is a completely valid question. This is why setting up the separate GWT account for the dev.domain.ext vs www.domain.ext. When you submit the removal request it will only be in the dev.domain.ext account.
The only thing you want to watch is that if you setup robots.txt in your dev environment you want to make sure that it does not get pushed out to your production server. That is the only gotcha as I see it.
-
thanks !
as er my last question theres no risk of accidentally taking out the main site as part of this process ?
cheers
dan
-
Thanks so much for that great advice
just a bit worried about accidentally getting main site removed by accident, i take it so long as its a brand new GWT account for that specific subdomain then this cant happen ?
Cheers
Dan
-
Here is a Google documentation on how to use the GWT to remove a page/directory/site and then the interaction with robots.txt
http://googlewebmastercentral.blogspot.com/2010/03/url-removal-explained-part-i-urls.html
"In order for a directory or site-wide removal to be successful, the directory or site must be disallowed in the site's robots.txt file."
Side story. I once had a subdomain that I needed to take out, but I could not modify the robots.txt file properly (long story). So, we used the GWT tool and the meta noindex tag. It still worked, but I think that would only be a backup approach to the one suggested by the documentation.
-
Usually, this would be true that you would need to use the noindex tag to get things out of the SERPs and need to leave the robots.txt "open" to the crawlers. But when you are working with the remove URL tool in GWT,they rx that you then put the site in robots.txt to keep them out of it
The removal tool in GWT takes care of Google taking the URLs out and then the robots.txt keeps the bots from coming back. Just a different sequence than if you were to use the noindex meta.
-
If you create the GWT account for the dev site and you submit for removal, GWT requires that you either a) have the site blocked in robots.tx or have a noindex meta tag on the pages. Otherwise they will just crawl you again later and you are back in the index. See my post from earlier.
-
Short answer - no dev sites should be public to start with to anyone (let along Google et alia). The simplest way is to put an htacess password on all your dev sites. You can do a password per person in your company, or just one general one that everyone on the dev team shares.
If you do have a dev site in the Serps, the simplest way to get it out is to setup a GWT account for that subdomain and then e.g. dev.yourdomain.ext and then go into that account and request removal of all pages. You just leave the form blank for the page to be removed and it takes out the whole site. You then need a robots.txt on dev.yourdomain.ext (different from the www. version) that disallows all pages all crawlers - that or use the noindex meta tag on all page.
After about 1 month (or you see that the pages are all out of the serps), then I would put up a password on that entire site and be done with it. Key point, dont put the password up until you let google try to spider and it sees the robots etc.
Also, if you have any other staging sites that are out there like test.yourdomain.ext etc. If they are not indexed, go ahead and put the password up on them to limit your exposure.
Public dev sites are the fastest way to get duplicate content into the index and to jack with the ranking of your current site. It is key that all of them are locked down. If one of your developers say it is no big deal, call BS, it is a big deal and it can cause a big mess.
-
Hey Dan,
In this case, I would not exclude crawling via robots.txt. Perhaps later after you have verified the URLs are out of the index.
Just because Google can't crawl a page, doesn't mean they won't keep it in the index. Excluding crawling will not get a page out of the index.
Add the NOINDEX, FOLLOW tag you listed above and give it some time.
Use GWT if it's urgent or the information is sensitive.
-
Thanks Anthony,
The staging area already exists and is indexable as far as i can tell
So i need to tell developers to exclude crawling via robots.txt, add a no-index tag to head of each page but keep it followed so still crawlable i.e. within the Head section of every page on the dev area
OR alternatively just remove urls from GWT)
If excluding crawling via robots.txt file then why do you need to add a noindex tag to each page too, surely the robots.txt deals with this situation ?
cheers
dan
-
Ideally when creating a new staging area, you'd want to exclude crawling via robots.txt.
Add the NoIndex tag to the head of your pages to get them removed from the SERPs. Make sure the page is still crawlable though, as if you exclude it in robots.txt first and then NoIndex it, Google won't be able to see the new NoIndex tag.
If there are not a lot of pages to remove, you can request page removal within Google Webmaster Tools.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
URLs dropping from index (Crawled, currently not indexed)
I've noticed that some of our URLs have recently dropped completely out of Google's index. When carrying out a URL inspection in GSC, it comes up with 'Crawled, currently not indexed'. Strangely, I've also noticed that under referring page it says 'None detected', which is definitely not the case. I wonder if it could be something to do with the following? https://www.seroundtable.com/google-ranking-index-drop-30192.html - It seems to be a bug affecting quite a few people. Here are a few examples of the URLs that have gone missing: https://www.ihasco.co.uk/courses/detail/sexual-harassment-awareness-training https://www.ihasco.co.uk/courses/detail/conflict-resolution-training https://www.ihasco.co.uk/courses/detail/prevent-duty-training Any help here would be massively appreciated!
Technical SEO | | iHasco0 -
Pages being flagged in Search Console as having a "no-index" tag, do not have a meta robots tag??
Hi, I am running a technical audit on a site which is causing me a few issues. The site is small and awkwardly built using lots of JS, animations and dynamic URL extensions (bit of a nightmare). I can see that it has only 5 pages being indexed in Google despite having over 25 pages submitted to Google via the sitemap in Search Console. The beta Search Console is telling me that there are 23 Urls marked with a 'noindex' tag, however when i go to view the page source and check the code of these pages, there are no meta robots tags at all - I have also checked the robots.txt file. Also, both Screaming Frog and Deep Crawl tools are failing to pick up these urls so i am a bit of a loss about how to find out whats going on. Inevitably i believe the creative agency who built the site had no idea about general website best practice, and that the dynamic url extensions may have something to do with the no-indexing. Any advice on this would be really appreciated. Are there any other ways of no-indexing pages which the dev / creative team might have implemented by accident? - What am i missing here? Thanks,
Technical SEO | | NickG-1230 -
No index tag robots.txt
Hi Mozzers, A client's website has a lot of internal directories defined as /node/*. I already added the rule 'Disallow: /node/*' to the robots.txt file to prevents bots from crawling these pages. However, the pages are already indexed and appear in the search results. In an article of Deepcrawl, they say you can simply add the rule 'Noindex: /node/*' to the robots.txt file, but other sources claim the only way is to add a noindex directive in the meta robots tag of every page. Can someone tell me which is the best way to prevent these pages from getting indexed? Small note: there are more than 100 pages. Thanks!
Technical SEO | | WeAreDigital_BE
Jens0 -
URLs with dashes between words or nothing at all? ( ../some-content vs. ../somecontent)
Here's a quick and easy question: Is there any problem with not using dashes in between words for URLs? Obviously the readability factor is a concern, but from a search engine standpoint? Thanks in advance!
Technical SEO | | tbinga0 -
New Page Showing Up On My Reports w/o Page Title, Words, etc - However, I didn't create it
I have a WordPress site and I was doing a crawl for errors and it is now showing up as of today that this page : https://thinkbiglearnsmart.com/event-registration/?event_id=551&name_of_event=HTML5 CSS3 is new and has no page title, words, etc. I am not even sure where this page or URL came from. I was messing with the robots.txt file to allow some /category/ posts that were being hidden, but I didn't re-allow anything with the above appendages. I just want to make sure that I didn't screw something up that is now going to impact my rankings - this was just a really odd message to come up as I didn't create this page recently - and that shouldnt even be a page accessible to the public. When I edit the page - it is using an Event Espresso (WordPress plugin) shortcode - and I don't want to noindex this page as it is all of my events. Sorry this post is confusing, any help or insight would be appreciated! I am also interested in hiring someone for some hourly consulting work on SEO type issues if anyone has any references. Thank you!
Technical SEO | | webbmason0 -
Can you have a /sitemap.xml and /sitemap.html on the same site?
Thanks in advance for any responses; we really appreciate the expertise of the SEOmoz community! My question: Since the file extensions are different, can a site have both a /sitemap.xml and /sitemap.html both siting at the root domain? For example, we've already put the html sitemap in place here: https://www.pioneermilitaryloans.com/sitemap Now, we're considering adding an XML sitemap. I know standard practice is to load it at the root (www.example.com/sitemap.xml), but am wondering if this will cause conflicts. I've been unable to find this topic addressed anywhere, or any real-life examples of sites currently doing this. What do you think?
Technical SEO | | PioneerServices0 -
No indexing url including query string with Robots txt
Dear all, how can I block url/pages with query strings like page.html?dir=asc&order=name with robots txt? Thanks!
Technical SEO | | HMK-NL0 -
What does it mean by 'blocked by Meta Robot'? How do I fix this?
When i get my crawl diagnostics, I am getting a blocked by Meta Robot, which means that my page is not being indexed in the search engines... obviously this is a major issue for organic traffic!!! What does it actually mean, and how can i fix it?
Technical SEO | | rolls1230