GWT URL Removal Tool Risky to Use for Duplicate Pages?
-
I was planning to remove lots of URL's via GWT that are highly duplicate alike pages (similar pages exist on other websites across the web). However, this Google article had me a bit concerned: https://support.google.com/webmasters/answer/1269119?hl=en
I already have "noindex, follow" on the pages I want to remove from the index, but Google seems to take ages to remove pages from index, which appear to drag down unique content pages from my site.
-
Hi
I have used the URL removal tool in the past to remove URLs with success - as we know it helps speed things up. What you have done is right and if you are patient Google will start removing each page as it crawls it again. You might find this confirmation from Google reassuring in your situation - https://support.google.com/webmasters/answer/93710?hl=en
Reading the article you posted of when not to use the tool I can't see that your pages fall into any of these categories - but either way I personally can't see using it causing an issue to be honest but its your call.
-
adding "nofollow" as well makes it even easier to get out of the index?
-
Last time I used URL removal in GWT was a long time ago and at that time the URL will not get out of the index for ever but for 90 days only and after that it will come again.
The better idea in any case is to use no index, no follow tag on the pages that you want out from the Google’s index!
Hope this helps!
-
issue with having pages that are similar to pages on other websites is the ratio of unique vs duplicate content is low and that can drag down other more unique pages ranking. The pages I have without much unique content is what users want: http://www.honoluluhi5.com/oahu/honolulu/hawaii-kai-homes/ but since content isn't unique I - unfortunately - need to noindex those pages and instead rank for this type of page: http://www.honoluluhi5.com/hawaii-kai-homes-real-estate/
When a user is looking for "….for sale" type keyword they want that first URL. Not the 2nd URL with pictures and video and writing. The "noindex, follow" is on the 1st URL, but still indexed after 1 month. I want to get de-indexed and I am trying to establish the risk associated with using that GWT tool - based on the article where G seems to indicate one shouldn't so easily use that tool. Conclusion is probably I have to be patient and wait for G to noindex those pages of mine. I look forward to the day G's algorithm can see the layout of a page and understand the value for users, even though it lacks unique content….
-
There's really no problem with having pages on your site having content that can be seen on other sites. Since you have noindexed them already, it shouldn't be a problem.
If they aren't really getting any traffic for you or aren't really bringing in anything that helps the site overall, then just take them off.
Focus on your new and existing content.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is it best to 301 redirect or use canonical Url when consolidating two pages?
I have build several pages (A and B) with high quantity content. Page A is aged and gets lots of organic traffic, ranks for lots of valuable keywords, and has only internal links to this page. Page B is newer (6 months) and gets little traffic, ranks for no keywords, but has terrific content and many high value external links. As Page A and B are related to a similar theme, I was going to merge content from page B onto page A, but don't know which would be the best approach for handling the links going to page B. For the purposes of keep as much link equity as possible, is it best to us a 301 redirect from B to A or use a canonical URL from B to A?
Intermediate & Advanced SEO | | Cutopia0 -
Duplicate URL Parameters for Blog Articles
Hi there, I'm working on a site which is using parameter URLs for category pages that list blog articles. The content on these pages constantly change as new posts are frequently added, the category maybe for 'Heath Articles' and list 10 blog posts (snippets from the blog). The URL could appear like so with filtering: www.domain.com/blog/articles/?taxonomy=health-articles&taxon=general www.domain.com/blog/articles/?taxonomy=health-articles&taxon=general&year=2016 www.domain.com/blog/articles/?taxonomy=health-articles&taxon=general&year=2016&page=1 All pages currently have the same Meta title and descriptions due to limitations with the CMS, they are also not in our xml sitemap I don't believe we should be focusing on ranking for these pages as the content on here are from blog posts (which we do want to rank for on the individual post) but there are 3000 duplicates and they need to be fixed. Below are the options we have so far: Canonical URLs Have all parameter pages within the category canonicalize to www.domain.com/blog/articles/?taxonomy=health-articles&taxon=general and generate dynamic page titles (I know its a good idea to use parameter pages in canonical URLs). WMT Parameter tool Tell Google all extra parameter tags belong to the main pages (e.g. www.domain.com/blog/articles/?taxonomy=health-articles&taxon=general&year=2016&page=3 belongs to www.domain.com/blog/articles/?taxonomy=health-articles&taxon=general). Noindex Remove all the blog category pages, I don't know how Google would react if we were to remove 3000 pages from our index (we have roughly 1700 unique pages) We are very limited with what we can do to these pages, if anyone has any feedback suggestions it would be much appreciated. Thanks!
Intermediate & Advanced SEO | | Xtend-Life0 -
Cached Alternate URL appearing as base page
Hi there, I'm currently targeting Australia and the US for one of my web-pages. One of my web-pages begin with a subdomain (au.site.com) and the other one is just the root domain (site.com). After searching the website on Australian Google and checking the description and title, it keeps the US ones (i.e. root domain) and after checking the cached copy, it was cached earlier today but it is displayed exactly as the American website when it is supposed to be the Australian one? In the url for the caching it appears as au.site.com while displaying the American page's content. Any ideas why? Thanks, Oliver
Intermediate & Advanced SEO | | oliverkuchies0 -
Canonical use when dynamically placing items on "all products" page
Hi all, We're trying to get our canonical situation straightened out. We have a section of our site with 100 product pages in it (in our case a city with hotels that we've reviewed), and we have a single page where we list them all out--an "all products" page called "all.html." However, because we have 100 and that's a lot for a user to see at once, we plan to first show only 50 on "all.html." When the user scrolls down to the bottom, we use AJAX to place another 50 on the page (these come from another page called "more.html" and are placed onto "all.html"). So, as you scroll down from the front end, you see "all.html" with 100 listings. We have other listings pages that are sorted and filtered subsets of this list with little or no unique content. Thus, we want to place a canonical on those pages. Question: Should the canonical point to "all.html"? Would spiders get confused, because they see that all.html is only half the listings? Is it dangerous to dynamically place content on a page that's used as a canonical? Is this a non-issue? Thanks, Tom
Intermediate & Advanced SEO | | TomNYC0 -
Duplicate Page Content / Titles Help
Hi guys, My SEOmoz crawl diagnostics throw up thousands of Dup Page Content / Title errors which are mostly from the forum attached to my website. In-particular it's the forum user's profiles that are causing the issue, below is a sample of the URLs that are being penalised: http://www.mywebsite.com/subfolder/myforum/pop_profile.asp?mode=display&id=1308 I thought that by adding - http://www.mywebsite.com/subfolder/myforum/pop_profile.asp to my robots.txt file under 'Ignore' would cause the bots to overlook the thousands of profile pages but the latest SEOmoz crawl still picks them up. My question is, how can I get the bots to ignore these profile pages (they don't contain any useful content) and how much will this be affecting my rankings (bearing in mind I have thousands of errors for dup content and dup page titles). Thanks guys Gareth
Intermediate & Advanced SEO | | gaz33420 -
Add URL parameters in SEOMoz as per GWT?
Hi, this may be a tall order, or maybe it's already in place and I'm behind the times! Any chance on getting something like this going? Even handier, have SEOMoz import these settings directly from GWT. The issue comes into play when looking at my duplicate page content reports; I'm guessing that SEOMoz will continue showing these as duplicates even after I have tweaked GWT to read them properly. Haven't tested this theory as I just started down this road on GWT myself. Thanks. 🙂
Intermediate & Advanced SEO | | ntcma0 -
Is there a tool that lists all external followed URLs?
Is there a tool that lists all external followed URLs? Or maybe separates nofollowed and followed external URLs?
Intermediate & Advanced SEO | | MangoMan160 -
How are pages ranked when using Google's "site:" operator?
Hi, If you perform a Google search like site:seomoz.org, how are the pages displayed sorted/ranked? Thanks!
Intermediate & Advanced SEO | | anthematic0