Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How to remove all sandbox test site link indexed by google?
-
When develop site, I have a test domain is sandbox.abc.com, this site contents are same as abc.com.
But, now I search site:sandbox.abc.com and aware of content duplicate with main site abc.com
My question is how to remove all this link from goolge.
p/s: I have just add robots.txt to sandbox and disallow all pages.
Thanks,
-
Thanks Chirag Nirmal,
I have redirected 301 successfully. This is great solution.
Thank you!
-
Hmmm.. I think its the IP address of your computer i.e. goto http://whatismyipaddress.com/ and get your IP address and use it below.
RewriteCond %{REMOTE_HOST} !^123.45.67.8
RewriteRule (.*) http://www.vietnamvisacorp.com/$1 [R=301,L]I am sorry, not much of an Apache expert.
-
Thanks for you help?
Is the IP address you mentioned is my localhost? i.e. My sandbox IP address is 123.45.67.8
The redirect 301 should be:RewriteCond %{REMOTE_ADDR} ^123.45.67.8
RewriteCond %{HTTP_HOST} !^$
RewriteRule /* http://www.vietnamvisacorp.com/ [R,L]or
RewriteCond %{REMOTE_HOST} !^123.45.67.8
RewriteRule (.*) http://www.vietnamvisacorp.com/$1 [R=301,L]Am I right?
Thanks,
-
Please see https://support.google.com/webmasters/bin/answer.py?hl=en&answer=47334
You can demote up to 100 URLs, and demotions are effective for 90 days from your most recent visit to the Sitelinks page in Webmaster Tools.
If I were you, then on http://sandbox.vietnamvisacorp.com/ I would put a 301 redirect based on ip address on all pages.
This will redirect all users to http://www.vietnamvisacorp.com/ based on ip address, i.e. if IP address = your ip then don't redirect otherwise redirect.
-
Thank you my friends. I could not understand denote a site link.
I want to denote all site link in site:http://sandbox.vietnamvisacorp.com/
How to denoted all in this:
-
For For this search result field: ???
-
For Demote this sitelink URL field: ???
-
-
If you are webmaster, you can demote the unwanted content using google web master tools.
Configuration -> Sitelinks-> Demote this sitelink URL:
And obviously, keep updating robots.txt moving forward.
-
"p/s: I have just add robots.txt to sandbox and disallow all pages."
That will do it, just have to wait till google picks up on those instructions (might take a while)
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Vanity URLs are being indexed in Google
We are currently using vanity URLs to track offline marketing, the vanity URL is structured as www.clientdomain.com/publication, this URL then is 302 redirected to the actual URL on the website not a custom landing page. The resulting redirected URL looks like: www.clientdomain.com/xyzpage?utm_source=print&utm_medium=print&utm_campaign=printcampaign. We have started to notice that some of the vanity URLs are being indexed in Google search. To prevent this from happening should we be using a 301 redirect instead of a 302 and will the Google index ignore the utm parameters in the URL that is being 301 redirect to? If not, any suggestions on how to handle? Thanks,
Technical SEO | | seogirl221 -
Removed Subdomain Sites Still in Google Index
Hey guys, I've got kind of a strange situation going on and I can't seem to find it addressed anywhere. I have a site that at one point had several development sites set up at subdomains. Those sites have since launched on their own domains, but the subdomain sites are still showing up in the Google index. However, if you look at the cached version of pages on these non-existent subdomains, it lists the NEW url, not the dev one in the little blurb that says "This is Google's cached version of www.correcturl.com." Clearly Google recognizes that the content resides at the new location, so how come the old pages are still in the index? Attempting to visit one of them gives a "Server Not Found" error, so they are definitely gone. This is happening to a couple of sites, one that was launched over a year ago so it doesn't appear to be a "wait and see" solution. Any suggestions would be a huge help. Thanks!!
Technical SEO | | SarahLK0 -
Fake Links indexing in google
Hello everyone, I have an interesting situation occurring here, and hoping maybe someone here has seen something of this nature or be able to offer some sort of advice. So, we recently installed a wordpress to a subdomain for our business and have been blogging through it. We added the google webmaster tools meta tag and I've noticed an increase in 404 links. I brought this up to or server admin, and he verified that there were a lot of ip's pinging our server looking for these links that don't exist. We've combed through our server files and nothing seems to be compromised. Today, we noticed that when you do site:ourdomain.com into google the subdomain with wordpress shows hundreds of these fake links, that when you visit them, return a 404 page. Just curious if anyone has seen anything like this, what it may be, how we can stop it, could it negatively impact us in anyway? Should we even worry about it? Here's the link to the google results. https://www.google.com/search?q=site%3Amshowells.com&oq=site%3A&aqs=chrome.0.69i59j69i57j69i58.1905j0j1&sourceid=chrome&es_sm=91&ie=UTF-8 (odd links show up on pages 2-3+)
Technical SEO | | mshowells0 -
Using the Google Remove URL Tool to remove https pages
I have found a way to get a list of 'some' of my 180,000+ garbage URLs now, and I'm going through the tedious task of using the URL removal tool to put them in one at a time. Between that and my robots.txt file and the URL Parameters, I'm hoping to see some change each week. I have noticed when I put URL's starting with https:// in to the removal tool, it adds the http:// main URL at the front. For example, I add to the removal tool:- https://www.mydomain.com/blah.html?search_garbage_url_addition On the confirmation page, the URL actually shows as:- http://www.mydomain.com/https://www.mydomain.com/blah.html?search_garbage_url_addition I don't want to accidentally remove my main URL or cause problems. Is this the right way this should look? AND PART 2 OF MY QUESTION If you see the search description in Google for a page you want removed that says the following in the SERP results, should I still go to the trouble of putting in the removal request? www.domain.com/url.html?xsearch_... A description for this result is not available because of this site's robots.txt – learn more.
Technical SEO | | sparrowdog1 -
Staging site and "live" site have both been indexed by Google
While creating a site we forgot to password protect the staging site while it was being built. Now that the site has been moved to the new domain, it has come to my attention that both the staging site (site.staging.com) and the "live" site (site.com) are both being indexed. What is the best way to solve this problem? I was thinking about adding a 301 redirect from the staging site to the live site via HTACCESS. Any recommendations?
Technical SEO | | melen0 -
Value of an embedded site vs. a direct link?
We have a new site that is a great resource for a serious subject (suicide). I have been getting many requests from various communities and clinics about help on embedding our site in their websites. Although I certainly don't want to keep this resource from being used as much as possible, I am curious about the SEO costs/benefit to having someone embed our site on their own website rather than provide a link to our website directly from theirs.
Technical SEO | | ron_adease1 -
How does Google Crawl Multi-Regional Sites?
I've been reading up on this on Webmaster Tools but just wanted to see if anyone could explain it a bit better. I have a website which is going live soon which is going to be set up to redirect to a localised URL based on the IP address i.e. NZ IP ranges will go to .co.nz, Aus IP addresses would go to .com.au and then USA or other non-specified IP addresses will go to the .com address. There is a single CMS installation for the website. Does this impact the way in which Google is able to search the site? Will all domains be crawled or just one? Any help would be great - thanks!
Technical SEO | | lemonz0 -
Removing robots.txt on WordPress site problem
Hi..am a little confused since I ticked the box in WordPress to allow search engines to now crawl my site (previously asked for them not to) but Google webmaster tools is telling me I still have robots.txt blocking them so am unable to submit the sitemap. Checked source code and the robots instruction has gone so a little lost. Any ideas please?
Technical SEO | | Wallander0