Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Staging website got indexed by google
-
Our staging website got indexed by google and now MOZ is showing all inbound links from staging site, how should i remove those links and make it no index.
Note- we already added Meta NOINDEX in head tag
-
Hi Dera Moz My Domain Is 18 Years Old But Da is don't increased i don't know why can you please help me and check my url cigars please check sir
#mozda
-
Its good that you already put the Meta NOINDEX.
Now, you can ask to remove the url of website from google index. Visit the google search console and request the url removal.
You can use the URL Removal Tool in Google Search Console to request the removal of specific URLs from Google's index.
To use the URL Removal Tool, you can:
- Open the Removals tool.
- Select the Temporary Removals tab.
- Click New Request.
- Select Next to complete the process.
Warm Regards
Rahul Gupta
Suvidit Academy -
Sydney's Best Chauffeur Car Service | A1 Corporate Cars Au
Sydney's Best Chauffeur Car Service is a premier provider of corporate chauffeured cars in Sydney, Australia. We offer top-of [url=https://a1corporatecars.com.au/]corporate cars Australia[/url] transportation solutions for business professionals, executives, and VIP clients who demand the highest service and comfort. With a fleet of luxury vehicles and experienced professional chauffeurs, we ensure a seamless and luxurious travel experience for our esteemed customers.
-
If your staging website has been indexed by Google, it means that Google's web crawlers have discovered and added your staging site's pages to their search index. This is typically not desirable because staging websites are meant for testing and development purposes and often contain incomplete or confidential content.
To address this issue, you can take several steps. Firstly, ensure that your staging website has a "robots.txt" file configured properly. This file tells search engines which parts of your website to crawl and index. In the case of a staging site, you can disallow all web crawlers from indexing it by using a "robots.txt" file.
Another effective measure is to include a "noindex" meta tag in the HTML of your staging website's pages. This tag instructs search engines not to index the page, adding an extra layer of protection.
Consider password-protecting your staging website using HTTP authentication. This adds an additional layer of security and ensures that only authorized users can access the site.
To further mitigate indexing issues, you can set up your staging website on a subdomain or a subdirectory instead of a separate domain. Google is less likely to index staging content if it's located in a subdomain or subdirectory.
If your staging site is already indexed, you can request the removal of specific URLs from Google's index using the Google Search Console's URL Removal Tool. This is a more proactive approach to remove already indexed content.
Lastly, regularly monitor your staging website to ensure it remains hidden from search engines and that any changes to the robots.txt file or meta tags are being followed. It's a good practice to implement these measures before you create or launch a staging website to prevent it from being indexed in the first place.
Remember that it may take some time for Google to update its index and remove your staging site's pages. Be patient and continue to monitor the situation closely to ensure the desired results are achieved.
-
If a staging website (a non-production or testing version) gets indexed by Google, it can lead to privacy, user experience, and SEO issues. To address this, use methods like robots.txt, "noindex" meta tags, or password protection to prevent indexing. If already indexed, request removal through Google Search Console to ensure only the production site is visible in search results.
-
If your staging website has been indexed by Google, it means that Google's search engine has discovered and included your staging site in its search results. This is not an ideal situation since staging websites are usually intended for testing and development purposes, and you may not want them publicly accessible.
To address this issue, you can take a few steps:
Use a robots.txt file: Create a robots.txt file on your staging website and instruct search engines not to index it. This file specifies which areas of your site search engines should or should not crawl.
Add a noindex meta tag: Insert a "noindex" meta tag in the head section of your staging website's HTML. This tag tells search engines not to index that specific page.
Password protect your staging website: Implement password protection on your staging environment to ensure that only authorized users can access it. This can be done through various authentication methods, depending on your setup.
Remember that these steps can help prevent further indexing, but they may not immediately remove your staging site from the search results. It might take some time for search engines to re-crawl your site and recognize the changes you made.
-
If your staging website gets indexed by Google, you should take these steps:
( Atlantic Immigration Pilot Program application form)
Use a robots.txt file to disallow indexing.
Request removal of indexed pages via Google Search Console.
Canada PR
Add a "noindex, nofollow" meta tag to staging pages.
Consider password protecting the staging site.
Ensure canonical URLs point to the production site.
These actions will help prevent your incomplete or sensitive staging content from appearing in Google search results.
Best digital marketing agency -
If your staging website has been indexed by Google, it means that Google's search engine has crawled and added your staging site's pages to its search index. This is typically not desired because staging websites are not meant for public access and may contain incomplete or sensitive content.
To address this issue, you should take the following steps:
Disallow indexing: Use a robots.txt file to instruct search engines not to crawl and index your staging website. You can add the following lines to your robots.txt file to disallow all search engines:
makefile
Copy code
User-agent: *
Disallow: /
Place this robots.txt file in the root directory of your staging website.Remove indexed pages: You can request Google to remove indexed pages from its search results by using the Google Search Console's "Remove URLs" tool. Log in to your Google Search Console account, select your property, go to the "Index" section, and choose "Removals." From there, you can temporarily hide specific URLs from Google search results.
Use noindex meta tags: On your staging website's pages, you can add a meta tag to indicate that the page should not be indexed. Add the following meta tag within the HTML <head> section of each page you want to exclude:
html
Copy code
<meta name="robots" content="noindex, nofollow">
This tag tells search engines not to index the page or follow any links on it.Password protection: Consider adding password protection to your staging website, so only authorized users can access it. This adds an additional layer of security and privacy.
Update canonical URLs: Ensure that your staging website's canonical URLs (if used) point to the production website, not the staging one. This helps search engines understand the preferred version of your content.
After taking these steps, monitor your staging website to ensure it's no longer being indexed by Google. Keep in mind that it may take some time for changes to take effect and for Google to de-index your staging content.
-
@Asmi-Ta said in Staging website got indexed by google:
Our staging website got indexed by google and now MOZ is showing all inbound links from staging site, how should i remove those links and make it no index.
Note- we already added Meta NOINDEX in head tagTo remove indexed staging site links and prevent further indexing, take these steps: Add a "Disallow" rule for the staging site in your
robots.txt
file, use 301 redirects for indexed staging URLs to point to production, update all internal links to production URLs, request URL removals through Google Search Console's "Fetch as Google" and URL Removal Tool, submit an updated production sitemap, and monitor Google Search Console for updates. Be patient, as it may take time for search engines to de-index staging URLs and re-crawl your site. Ensure the staging site has a "noindex" tag in its<head>
section.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Advise on the right way to block country specific users but not block Googlebot - and not be seen to be cloaking. Help please!
Hi, I am working on the SEO of an online gaming platform - a platform that can only be accessed by people in certain countries, where the games and content are legally allowed.
International SEO | | MarkCanning
Example: The games are not allowed in the USA, but they are allowed in Canada. Present Situation:
Presently when a user from the USA visits the site they get directed to a restricted location page with the following message: RESTRICTED LOCATION
Due to licensing restrictions, we can't currently offer our services in your location. We're working hard to expand our reach, so stay tuned for updates! Because USA visitors are blocked Google which primarily (but not always) crawls from the USA is also blocked, so the company webpages are not being crawled and indexed. Objective / What we want to achieve: The website will have multiple region and language locations. Some of these will exist as standalone websites and others will exist as folders on the domain. Examples below:
domain.com/en-ca [English Canada]
domain.com/fr-ca [french Canada]
domain.com/es-mx [spanish mexico]
domain.com/pt-br [portugese brazil]
domain.co.in/hi [hindi India] If a user from USA or another restricted location tries to access our site they should not have access but should get a restricted access message.
However we still want google to be able to access, crawl and index our pages. Can i suggest how do we do this without getting done for cloaking etc? Would this approach be ok? (please see below) We continue to work as the present situation is presently doing, showing visitors from the USA a restricted message.
However rather than redirecting these visitors to a restricted location page, we just black out the page and show them a floating message as if it were a model window.
While Googlebot would be allowed to visit and crawl the website. I have also read that it would be good to put paywall schema on each webpage to let Google know that we are not cloaking and its a restricted paid page. All public pages are accessible but only if the visitor is from a location that is not restricted Any feedback and direction that can be given would be greatly appreciated as i am new to this angle of SEO. Sincere thanks,0 -
Who is correct - please help!
I have a website with a lot of product pages - often thousands of pages. As each of these pages is for a specific lease car they are often only fractionally different from other pages. The urls are too long, the H1 is often too long and the Title is often too long for "SEO best practice". And they do create duplication issues according to MOZ. Some people tell me to change them to noindex/nofollow whilst others tell me to leave them as they are as best not to hide from google crawler. Any advice will be gratefully received. Thanks for listening.
Technical SEO | | jlhitch0 -
How do internal search results get indexed by Google?
Hi all, Most of the URLs that are created by using the internal search function of a website/web shop shouldn't be indexed since they create duplicate content or waste crawl budget. The standard way to go is to 'noindex, follow' these pages or sometimes to use robots.txt to disallow crawling of these pages. The first question I have is how these pages actually would get indexed in the first place if you wouldn't use one of the options above. Crawlers follow links to index a website's pages. If a random visitor comes to your site and uses the search function, this creates a URL. There are no links leading to this URL, it is not in a sitemap, it can't be found through navigating on the website,... so how can search engines index these URLs that were generated by using an internal search function? Second question: let's say somebody embeds a link on his website pointing to a URL from your website that was created by an internal search. Now let's assume you used robots.txt to make sure these URLs weren't indexed. This means Google won't even crawl those pages. Is it possible then that the link that was used on another website will show an empty page after a while, since Google doesn't even crawl this page? Thanks for your thoughts guys.
Intermediate & Advanced SEO | | Mat_C0 -
Google Is Indexing my 301 Redirects to Other sites
Long story but now i have a few links from my site 301 redirecting to youtube videos or eCommerce stores. They carry a considerable amount of traffic that i benefit from so i can't take them down, and that traffic is people from other websites, so basically i have backlinks from places that i don't own, to my redirect urls (Ex. http://example.com/redirect) My problem is that google is indexing them and doesn't let them go, i have tried blocking that url from robots.txt but google is still indexing it uncrawled, i have also tried allowing google to crawl it and adding noindex from robots.txt, i have tried removing it from GWT but it pops back again after a few days. Any ideas? Thanks!
Intermediate & Advanced SEO | | cuarto7150 -
How can I get a list of every url of a site in Google's index?
I work on a site that has almost 20,000 urls in its site map. Google WMT claims 28,000 indexed and a search on Google shows 33,000. I'd like to find what the difference is. Is there a way to get an excel sheet with every url Google has indexed for a site? Thanks... Mike
Intermediate & Advanced SEO | | 945010 -
How is Google crawling and indexing this directory listing?
We have three Directory Listing pages that are being indexed by Google: http://www.ccisolutions.com/StoreFront/jsp/ http://www.ccisolutions.com/StoreFront/jsp/html/ http://www.ccisolutions.com/StoreFront/jsp/pdf/ How and why is Googlebot crawling and indexing these pages? Nothing else links to them (although the /jsp.html/ and /jsp/pdf/ both link back to /jsp/). They aren't disallowed in our robots.txt file and I understand that this could be why. If we add them to our robots.txt file and disallow, will this prevent Googlebot from crawling and indexing those Directory Listing pages without prohibiting them from crawling and indexing the content that resides there which is used to populate pages on our site? Having these pages indexed in Google is causing a myriad of issues, not the least of which is duplicate content. For example, this file <tt>CCI-SALES-STAFF.HTML</tt> (which appears on this Directory Listing referenced above - http://www.ccisolutions.com/StoreFront/jsp/html/) clicks through to this Web page: http://www.ccisolutions.com/StoreFront/jsp/html/CCI-SALES-STAFF.HTML This page is indexed in Google and we don't want it to be. But so is the actual page where we intended the content contained in that file to display: http://www.ccisolutions.com/StoreFront/category/meet-our-sales-staff As you can see, this results in duplicate content problems. Is there a way to disallow Googlebot from crawling that Directory Listing page, and, provided that we have this URL in our sitemap: http://www.ccisolutions.com/StoreFront/category/meet-our-sales-staff, solve the duplicate content issue as a result? For example: Disallow: /StoreFront/jsp/ Disallow: /StoreFront/jsp/html/ Disallow: /StoreFront/jsp/pdf/ Can we do this without risking blocking Googlebot from content we do want crawled and indexed? Many thanks in advance for any and all help on this one!
Intermediate & Advanced SEO | | danatanseo0 -
Our login pages are being indexed by Google - How do you remove them?
Each of our login pages show up under different subdomains of our website. Currently these are accessible by Google which is a huge competitive advantage for our competitors looking for our client list. We've done a few things to try to rectify the problem: - No index/archive to each login page Robot.txt to all subdomains to block search engines gone into webmaster tools and added the subdomain of one of our bigger clients then requested to remove it from Google (This would be great to do for every subdomain but we have a LOT of clients and it would require tons of backend work to make this happen.) Other than the last option, is there something we can do that will remove subdomains from being viewed from search engines? We know the robots.txt are working since the message on search results say: "A description for this result is not available because of this site's robots.txt – learn more." But we'd like the whole link to disappear.. Any suggestions?
Intermediate & Advanced SEO | | desmond.liang1 -
XML Sitemap index within a XML sitemaps index
We have a similar problem to http://www.seomoz.org/q/can-a-xml-sitemap-index-point-to-other-sitemaps-indexes Can a XML sitemap index point to other sitemaps indexes? According to the "Unique Doll Clothing" example on this link, it seems possible http://www.seomoz.org/blog/multiple-xml-sitemaps-increased-indexation-and-traffic Can someone share an XML Sitemap index within a XML sitemaps index example? We are looking for the format to implement the same on our website.
Intermediate & Advanced SEO | | Lakshdeep0