Canonical from NOINDEX,FOLLOW pages - Bad idea?
-
Hi,
We have an extensive online shop in Magento - to ensure that some of the pages with query strings are not indexed, we implemented a conditional NOINDEX,FOLLOW so that it will stop indexing any pages that have querystrings on it -
We do need to use Canonical also - for other reasons - so my question is:
If you have a page that is NOINDEX,FOLLOW and it has a rel canonical pointing to original, would it transfer that NOINDEX,FOLLOW to the main original page causing it problems?
Thanks!
-
Hi there
No, the canonical will not pass the meta robots directive to the original page, so you're safe there.
What you're effectively doing is using two ways to prevent duplication - the canonical will instruct web crawlers not to index versions of the URL with query strings, just as the noindex,nofollow tags will.
Nothing wrong with using two methods simultaneously to do this - always a good idea to be safe - and so the end result will be that the URLs with query strings will be very, very unlikely to be indexed.
-
There was a very good article from Dr. Pete about HTTP status.
Canonicals do not transfer information like noindex, follow.
What they transfer is the incoming "link juice" to the original version of the page. So basically it counts as a redirect for Search Engines without redirecting the visitor, which means it won't be indexed (the non-original version) and all the link juice that the page got will be transferred to the original version.
I hope it helps,
Istvan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Need only tens of pages to be indexed out of hundreds: Robots.txt is Okay for Google to proceed with?
Hi all, We 2 sub domains with hundreds of pages where we need only 50 pages to get indexed which are important. Unfortunately the CMS of these sub domains is very old and not supporting "noindex" tag to be deployed on page level. So we are planning to block the entire sites from robots.txt and allow the 50 pages needed. But we are not sure if this is the right approach as Google been suggesting to depend mostly on "noindex" than robots.txt. Please suggest whether we can proceed with robots.txt file. Thanks
Algorithm Updates | | vtmoz0 -
Do the back-links go wasted when anchor text or context content doesn't match with page content?
Hi Community, I have seen number of back-links where the content in that link is not matching with page content. Like page A linking to page B, but content is not really relevant beside brand name. Like page with "vertigo tiles" linked to page about "vertigo paints" where "vertigo" is brand name. Will these kind of back-links completely get wasted? I have also found some broken links which I'm planning to redirect to existing pages just to reclaim the back-links even though the content relevancy is not much beside brand name. Are these back-links are beneficial or not? Thanks
Algorithm Updates | | vtmoz0 -
Big hit taken on Google Search in Jan - Any Ideas?
Hello, I manage a news site that gets new items posted daily. We had had a pretty even keel with Google search and ranking for some time now only on the 9th Jan we took a massive drop and have no recovered except for one big spike on the 29th January. The only think we had done differently was not post as much over Christmas for about a week as people were on holiday but if this was the reason for it the posting is back to normal now and has been since the 6th Jan and nothing has recovered. The site is wjlondon.com - any ideas greatly appreciated. Thank you
Algorithm Updates | | luwhosjack0 -
Shared Hosting - Bad for SEO? (exp. Godaddy)
There were a lot of questions and data on this a few years back and nothing terribly recent so i wanted to get the discussion going again and see if any new data has been published. Is hosting your website on a shared host like Godaddy or Network Solutions going to hurt your rankings because their holds a chance that you could be on the same IP as spammy websites? My gut feeling is no primarily because almost 90% of the worldwide web is on shared hosting but i do not have a lot of data to back it up. Id love to hear some feedback. Cheers - Kyle
Algorithm Updates | | kchandler0 -
Canonicalization on more than one page?
is it proper to "canocalize" more than one page in a site? Or should it only be on the home page? eg: http://www.sundayschoolnetwork.com">
Algorithm Updates | | sakeith0 -
Google has indexed a lot of test pages/junk from the development days.
With hind site I understand that this could have been avoided if robots.txt was configured properly. My website is www.clearvisas.com, and is indexed with both the www subdomain and with out. When I run site:clearvisas.com in Google I get 1,330 - All junk from the development days. But when I run site:www.clearvisas.com in Google I get 66 - these results all post development and more in line with what I wanted to be indexed. Will 1,330 junk pages hurt my seo? Is it possible to de-index them and should I? If the answer is yes to any of the questions how should I proceed? Kind regards, Fuad
Algorithm Updates | | Fuad_YK0 -
Stop google indexing CDN pages
Just when I thought I'd seen it all, google hits me with another nasty surprise! I have a CDN to deliver images, js and css to visitors around the world. I have no links to static HTML pages on the site, as far as I can tell, but someone else may have - perhaps a scraper site? Google has decided the static pages they were able to access through the CDN have more value than my real pages, and they seem to be slowly replacing my pages in the index with the static pages. Anyone got an idea on how to stop that? Obviously, I have no access to the static area, because it is in the CDN, so there is no way I know of that I can have a robots file there. It could be that I have to trash the CDN and change it to only allow the image directory, and maybe set up a separate CDN subdomain for content that only contains the JS and CSS? Have you seen this problem and beat it? (Of course the next thing is Roger might look at google results and start crawling them too, LOL) P.S. The reason I am not asking this question in the google forums is that others have asked this question many times and nobody at google has bothered to answer, over the past 5 months, and nobody who did try, gave an answer that was remotely useful. So I'm not really hopeful of anyone here having a solution either, but I expect this is my best bet because you guys are always willing to try.
Algorithm Updates | | loopyal0 -
Will google punish us for using formulaic keyword-rich content on different pages on our site?
We have 100 to 150 words of SEO text per page on www.storitz.com. Our challenge is that we are a storage property aggregator with hundreds of metros. We have to distinguish each city with relevant and umique text. If we use a modular approach where we mix and match pre-written (by us) content, demographic and location oriented text in an attempt to create relevant and unique text for multiple (hundreds) of pages on our site, will we be devalued by Google?
Algorithm Updates | | Storitz0