GWT URL Removal Tool Risky to Use for Duplicate Pages?
-
I was planning to remove lots of URL's via GWT that are highly duplicate alike pages (similar pages exist on other websites across the web). However, this Google article had me a bit concerned: https://support.google.com/webmasters/answer/1269119?hl=en
I already have "noindex, follow" on the pages I want to remove from the index, but Google seems to take ages to remove pages from index, which appear to drag down unique content pages from my site.
-
Hi
I have used the URL removal tool in the past to remove URLs with success - as we know it helps speed things up. What you have done is right and if you are patient Google will start removing each page as it crawls it again. You might find this confirmation from Google reassuring in your situation - https://support.google.com/webmasters/answer/93710?hl=en
Reading the article you posted of when not to use the tool I can't see that your pages fall into any of these categories - but either way I personally can't see using it causing an issue to be honest but its your call.
-
adding "nofollow" as well makes it even easier to get out of the index?
-
Last time I used URL removal in GWT was a long time ago and at that time the URL will not get out of the index for ever but for 90 days only and after that it will come again.
The better idea in any case is to use no index, no follow tag on the pages that you want out from the Google’s index!
Hope this helps!
-
issue with having pages that are similar to pages on other websites is the ratio of unique vs duplicate content is low and that can drag down other more unique pages ranking. The pages I have without much unique content is what users want: http://www.honoluluhi5.com/oahu/honolulu/hawaii-kai-homes/ but since content isn't unique I - unfortunately - need to noindex those pages and instead rank for this type of page: http://www.honoluluhi5.com/hawaii-kai-homes-real-estate/
When a user is looking for "….for sale" type keyword they want that first URL. Not the 2nd URL with pictures and video and writing. The "noindex, follow" is on the 1st URL, but still indexed after 1 month. I want to get de-indexed and I am trying to establish the risk associated with using that GWT tool - based on the article where G seems to indicate one shouldn't so easily use that tool. Conclusion is probably I have to be patient and wait for G to noindex those pages of mine. I look forward to the day G's algorithm can see the layout of a page and understand the value for users, even though it lacks unique content….
-
There's really no problem with having pages on your site having content that can be seen on other sites. Since you have noindexed them already, it shouldn't be a problem.
If they aren't really getting any traffic for you or aren't really bringing in anything that helps the site overall, then just take them off.
Focus on your new and existing content.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate page content errors for Web App Login
Hi There I have 6 duplicate content errors, but they are for the WebApp login from our website. I have put a Noindex on the Sitemap to stop google from indexing them to see if that would work. But it didn't. These links as far as I can see are not even on the website www.skemaz.net, but are links beyond the website and on the Web App itself eg : <colgroup><col width="529"></colgroup>
Intermediate & Advanced SEO | | Skemazer
| http://login.skemaz.net |
| http://login.skemaz.net/LogIn?ReturnUrl=%2Fchangepassword |
| http://login.skemaz.net/Login |
| http://login.skemaz.net/LogIn?ReturnUrl=%2FHome | Any suggestions would be greatly appreciated. Kind regards Sarah0 -
Removing pages from index
My client is running 4 websites on ModX CMS and using the same database for all the sites. Roger has discovered that one of the sites has 2050 302 redirects pointing to the clients other sites. The Sitemap for the site in question includes 860 pages. Google Webmaster Tools has indexed 540 pages. Roger has discovered 5200 pages and a Site: query of Google reveals 7200 pages. Diving into the SERP results many of the pages indexed are pointing to the other 3 sites. I believe there is a configuration problem with the site because the other sites when crawled do not have a huge volume of redirects. My concern is how can we remove from Google's index the 2050 pages that are redirecting to the other sites via a 302 redirect?
Intermediate & Advanced SEO | | tinbum0 -
How long takes to a page show up in Google results after removing noindex from a page?
Hi folks, A client of mine created a new page and used meta robots noindex to not show the page while they are not ready to launch it. The problem is that somehow Google "crawled" the page and now, after removing the meta robots noindex, the page does not show up in the results. We've tried to crawl it using Fetch as Googlebot, and then submit it using the button that appears. We've included the page in sitemap.xml and also used the old Google submit new page URL https://www.google.com/webmasters/tools/submit-url Does anyone know how long will it take for Google to show the page AFTER removing meta robots noindex from the page? Any reliable references of the statement? I did not find any Google video/post about this. I know that in some days it will appear but I'd like to have a good reference for the future. Thanks.
Intermediate & Advanced SEO | | fabioricotta-840380 -
HELP! How does one prevent regional pages as being counted as "duplicate content," "duplicate meta descriptions," et cetera...?
The organization I am working with has multiple versions of its website geared towards the different regions. US - http://www.orionhealth.com/ CA - http://www.orionhealth.com/ca/ DE - http://www.orionhealth.com/de/ UK - http://www.orionhealth.com/uk/ AU - http://www.orionhealth.com/au/ NZ - http://www.orionhealth.com/nz/ Some of these sites have very similar pages which are registering as duplicate content, meta descriptions and titles. Two examples are: http://www.orionhealth.com/terms-and-conditions http://www.orionhealth.com/uk/terms-and-conditions Now even though the content is the same, the navigation is different since each region has different product options / services, so a redirect won't work since the navigation on the main US site is different from the navigation for the UK site. A rel=canonical seems like a viable option, but (correct me if I'm wrong) it tells search engines to only index the main page, in this case, it would be the US version, but I still want the UK site to appear to search engines. So what is the proper way of treating similar pages accross different regional directories? Any insight would be GREATLY appreciated! Thank you!
Intermediate & Advanced SEO | | Scratch_MM0 -
Using author on every page of website?
I'm currently get to grips with schema and one thing im using is author on my blog posts and seeing my photo etc on organic searches which are related. I see one of my competitors is using author on every page on their website, not just blog posts etc. Are there any recommendation when it should be used? Should it be site wide or is it really intended for blog posts etc? Would it be wrong for me to use on every page of my website as one of my businesses is myself as a lone person? This is what you get when searching for driving lessons in just about any town! https://www.google.co.uk/#gs_rn=15&gs_ri=psy-ab&tok=LS_DOrAHswmHC9_8AJZEJA&suggest=p&pq=driving instructor brighton&cp=20&gs_id=1k2&xhr=t&q=driving+lessons+crawley&es_nrs=true&pf=p&sclient=psy-ab&oq=driving+lessons+craw&gs_l=&pbx=1&bav=on.2,or.r_cp.r_qf.&bvm=bv.47244034,d.d2k&fp=45c2f917e11bca99&biw=1680&bih=843 Any comments welcome! Antony
Intermediate & Advanced SEO | | Ant710 -
Duplicate content throughout multiple URLs dilemma
We have a website with lots of categories and there are problems that some subcategories have identical content on them. So, is it enough to just add different text on those problematic subcategories or we need to use "canonical" tag to main category. Same dilemma is with our search system and duplicate content. For example, "/category/sports" URL would have similar to identical content with "/search/sports" and "/search/sports-fitness/" URLs. Ranking factors is important for all different categories and subcategories. Ranking factors is also important for search individual keywords. So, the question is, how to make them somehow unique/different to rank on all those pages well? Would love to hear advices how it can be solved using different methods and how it would affect our rankings. When we actually need to use "canonical" tag and when 301 redirect is better. Thanks!
Intermediate & Advanced SEO | | versliukai0 -
How do I fix the error duplicate page content and duplicate page title?
On my site www.millsheating.co.uk I have the error message as per the question title. The conflict is coming from these two pages which are effectively the same page: www.millsheating.co.uk www.millsheating.co.uk/index I have added a htaccess file to the root folder as I thought (hoped) it would fix the problem but I doesn't appear to have done so. this is the content of the htaccess file: Options +FollowSymLinks RewriteEngine On RewriteCond %{HTTP_HOST} ^millsheating.co.uk RewriteRule (.*) http://www.millsheating.co.uk/$1 [R=301,L] RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /index\.html\ HTTP/ RewriteRule ^index\.html$ http://www.millsheating.co.uk/ [R=301,L] AddType x-mapp-php5 .php
Intermediate & Advanced SEO | | JasonHegarty0 -
Removing URLs in bulk when directory exclusion isn't an option?
I had a bunch of URLs on my site that followed the form: http://www.example.com/abcdefg?q=&site_id=0000000048zfkf&l= There were several million pages, each associated with a different site_id. They weren't very useful, so we've removed them entirely and now return a 404.The problem is, they're still stuck in Google's index. I'd like to remove them manually, but how? There's no proper directory (i.e. /abcdefg/) to remove, since there's no trailing /, and removing them one by one isn't an option. Is there any other way to approach the problem or specify URLs in bulk? Any insights are much appreciated. Kurus
Intermediate & Advanced SEO | | kurus1