Best Practice Approaches to Canonicals vs. Indexing in Google Sitemap vs. No Follow Tags
-
Hi There,
I am working on the following website: https://wave.com.au/
I have become aware that there are different pages that are competing for the same keywords.
For example, I just started to update a core, category page - Anaesthetics (https://wave.com.au/job-specialties/anaesthetics/) to focus mainly around the keywords ‘Anaesthetist Jobs’.
But I have recognized that there are ongoing landing pages that contain pretty similar content:
We want to direct organic traffic to our core pages e.g. (https://wave.com.au/job-specialties/anaesthetics/).
This then leads me to have to deal with the duplicate pages with either a canonical link (content manageable) or maybe alternatively adding a no-follow tag or updating the robots.txt. Our resident developer also suggested that it might be good to use Google Index in the sitemap to tell Google that these are of less value?
What is the best approach? Should I add a canonical link to the landing pages pointing it to the category page? Or alternatively, should I use the Google Index? Or even another approach?
Any advice would be greatly appreciated.
Thanks!
-
This all sounds good, just make sure before you proceed, you use GA to check what % of your SEO (segment: "Organic") traffic comes from these URLs. Don't act on a hunch, act on data!
-
Thank you for the comprehensive response this is greatly appreciated my friend.
Yes, I agree. I have since read further and have completely ruled out blocking (robots txt. etc) as an option.
I went back and read some more Moz/SEO articles and I think I have narrowed it down to either:
a) canonicals pointing from the landing pages to the core website category pages
b) NoIndex/Follow tags on the landing pages
Basically, I think the key contextual factors to keep in mind are that:
- The landing pages are basically just sent to people directly by our recruiters in emails and over the phone, so they are almost counted as direct traffic.
- It just contains a form and doesn't encourage click through into our core website beside logo etc. - we just want them to register directly on that page.
- Over the past year, the visits on the landing pages were much, much less, and the bounce rate and exit % was higher.
- my manager has told me to prioritise the SEO towards the core category pages as they see the landing pages as purely for UX/registrations/useful to internal business recruiting practices rather than encouraging organic traffic.
I think canonicals would probably work the best since in some cases the landing pages were ranking higher than the category pages and it should hopefully transfer a bit of ranking power to the category pages.
But perhaps you are right and I can batch apply canonicals monitor the results and then progress.
Once again, thank you for your response.
-
First of all keep in mind that Google has chosen the pages it is deciding to rank for one reason or another, and that canonical tags do not consolidate link equity (SEO authority) in the same way which 301 redirects do
As such, it's possible that you could implement a very 'logical' canonical tag structure, but for whatever reason Google may not give your new 'canonical' URLs the same rankings which it ascribed to the old URLs. So there is a possibility here that, you could lose some rankings! Google's acceptance of both the canonical tag and the 301 redirect depends upon the (machine-like) similarity of the content on both URLs
Think of Boolean string similarity. You get two strings of text, whack them into a tool like this one - and it tells you the 'percentage' of similarity between the two text strings. Google operate something similar yet infinitely more sophisticated. No one has told me that they do this, I have observed it over hundreds of site migration projects where, sometimes Google gives the new site loads of SEO authority through the 301s and sometimes not much at all. For me, the two main causes of Google refusing to accept new canonical URLs are redirect chains (which could include soft redirect chains) but also content 'dissimilarity'. Basically, content has won links and interactions on one URL which prove it is popular and authoritative. If you move that content somewhere else, or tell Google to go somewhere else instead - they have to be pretty certain that the new content is pretty much the same, otherwise it's a risk to them and an 'unknown quantity' in the SERPs (in terms of CTR and stuff)
If you're pretty damn sure that you have loads of URLs which are essentially the same, read the same, reference the same prices for things (one isn't cheaper than the other), that Google has really chosen the wrong page to rank in terms of Google-user click-through UX, then go ahead and lay out your canonical tag strategy
Personally I'd pick sections of the site and do it one part at a time in isolation, so you can minimise losses from disturbing Google and also measure your efforts more effectively / efficiently
If you no-index and robots-block URLs, it KILLS their SEO authority (dead) instead of moving it elsewhere (so steer clear of those except in extreme situations, they're really a last resort if you have the worst sprawling architecture imaginable). 301 redirects can shift ranking URLs and relevance, but don't pipe much authority. 301 redirects (if handled correctly) do all three things
What you have to ask yourself is, if you flat out deleted the pages you don't want to rank (obviously you wouldn't do this, as it would cause internal UX issues on your site) - if you did that, would Google:
A) Rank the other pages in their place from your site, which you want Google to rank
B) Give up on you and just rank similar pages (to the ones you don't want to rank) from other, competing sites instead
If you think (A) - take a measured, sectioned, small approach to canonical tag deployment and really test it before full roll-out. If you think (B), then you are admitting that there's something more Google-friendly one the pages you don't want to be ranking and just have to accept - no, your Google->conversion funnel will never be completely perfect like how you want it to be. You have to satisfy Google, not the other way around
Hope that helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Why did Google cache & index a different domain than my own?
We own www.homemenorca.com, a real estate website based in Spain. Pages from this domain are not being indexed: https://www.google.com/search?q=site%3Awww.homemenorca.com&oq=site%3Awww.homemenorca.com&aqs=chrome..69i57j69i58j69i59l2.3504j0j7&sourceid=chrome&ie=UTF-8Please notice that the URLs are Home Menorca, but the titles are not Home Menorca, they are Fincas Mantolan, a completely different domain and company: http://www.fincasmantolan.com/. Furthermore, when we look at Google's cache of Home Menorca, we see a different website: http://webcache.googleusercontent.com/search?q=cache%3Awww.homemenorca.com%2Fen&oq=cache%3Awww.homemenorca.com%2Fen&aqs=chrome..69i57j69i58j69i59.1311j0j4&sourceid=chrome&ie=UTF-8We reviewed Google Search Console, Google Fetch, the canonical tags, the XML sitemap, and many more items. Google Search Console accepted our XML sitemap, but is only indexing 5-10% of the pages. Google is fetching and rendering the pages properly. However, we are not seeing the correct content being indexed in Google. We have seen issues with page loading times, loading content longer than 4 seconds, but are unsure why Google would be indexing a different domain.If you have suggestions or thoughts, we would very much appreciate it.Additional Language Issue:When a user searches "Home Menorca" from America or the UK with "English" selected in their browser as their default language, they are given a Spanish result. It seems to have accurate hreflang annotations within the head section on the HTML pages, but it is not working properly. Furthermore, Fincas Mantolan's search result is listed immediately below Home Menorca's Spanish result. We believe that if we fix the issue above, we will also fix the language issue. Please let us know any thoughts or recommendations that can help us. Thank you very much!
Intermediate & Advanced SEO | | CassG12340 -
Fetch as Google -- Does not result in pages getting indexed
I run a exotic pet website which currently has several types of species of reptiles. It has done well in SERP for the first couple of types of reptiles, but I am continuing to add new species and for each of these comes the task of getting ranked and I need to figure out the best process. We just released our 4th species, "reticulated pythons", about 2 weeks ago, and I made these pages public and in Webmaster tools did a "Fetch as Google" and index page and child pages for this page: http://www.morphmarket.com/c/reptiles/pythons/reticulated-pythons/index While Google immediately indexed the index page, it did not really index the couple of dozen pages linked from this page despite me checking the option to crawl child pages. I know this by two ways: first, in Google Webmaster Tools, if I look at Search Analytics and Pages filtered by "retic", there are only 2 listed. This at least tells me it's not showing these pages to users. More directly though, if I look at Google search for "site:morphmarket.com/c/reptiles/pythons/reticulated-pythons" there are only 7 pages indexed. More details -- I've tested at least one of these URLs with the robot checker and they are not blocked. The canonical values look right. I have not monkeyed really with Crawl URL Parameters. I do NOT have these pages listed in my sitemap, but in my experience Google didn't care a lot about that -- I previously had about 100 pages there and google didn't index some of them for more than 1 year. Google has indexed "105k" pages from my site so it is very happy to do so, apparently just not the ones I want (this large value is due to permutations of search parameters, something I think I've since improved with canonical, robots, etc). I may have some nofollow links to the same URLs but NOT on this page, so assuming nofollow has only local effects, this shouldn't matter. Any advice on what could be going wrong here. I really want Google to index the top couple of links on this page (home, index, stores, calculator) as well as the couple dozen gene/tag links below.
Intermediate & Advanced SEO | | jplehmann0 -
Google indexing pages from chrome history ?
We have pages that are not linked from site yet they are indexed in Google. It could be possible if Google got these pages from browser. Does Google takes data from chrome?
Intermediate & Advanced SEO | | vivekrathore0 -
Cross Domain Rel Canonical tags vs. Rel Canonical Tags for internal webpages
Today I noticed that one of my colleagues was pointing rel canonical tags to a third party domain on a few specific pages on a client's website. This was a standard rel canonical tag that was written Up to this point I haven't seen too many webmasters point a rel canonical to a third party domain. However after doing some reading in the Google Webmaster Tools blog I realized that cross domain rel canonicals are indeed a viable strategy to avoid duplicate content. My question is this; should rel canonical tags be written the same way when dealing with internal duplicate content vs. external duplicate content? Would a rel=author tag be more appropriate when addressing 3rd party website duplicate content issues? Any feedback would be appreciated.
Intermediate & Advanced SEO | | VanguardCommunications0 -
Canonical URL & sitemap URL mismatch
Hi We're running a Magento store which doesn't have too much stock rotation. We've implemented a plugin that will allow us to give products custom canonical URLs (basically including the category slug, which is not possible through vanilla Magento). The sitemap feature doesn't pick up on these URLs, so we're submitting URLs to Google that are available and will serve content, but actually point to a longer URL via a canonical meta tag. The content is available at each URL and is near identical (all apart from the breadcrumbs) All instances of the page point to the same canonical URL We are using the longer URL in our internal architecture/link building to show this preference My questions are; Will this harm our visibility? Aside from editing the sitemap, are there any other signals we could give Google? Thanks
Intermediate & Advanced SEO | | tomcraig860 -
When migrating website platforms but keeping the domain name how best do we add the new site to google webmaster tools? Best redirect practices?
We are moving from BigCommerce to Shopify but maintaining our domain name and need to make sure that all links redirect to their corresponding links. We understand the nature of 301s and are fine with that, but when it comes to adding the site to google webmaster tools, not losing link juice and the change of address tool we are kind of lost. Any advice would be most welcome. Thank you so much in advance!
Intermediate & Advanced SEO | | WNL0 -
Why would one of our section pages NOT be indexed by Google?
One of our higher traffic section pages is not being indexed by Google. The products that reside on this section page ARE indexed by Google and are on page 1. So why wouldn't the section page be even listed and indexed? The meta title is accurate, meta description is good. I haven't received any notices in Webmaster Tools. Is there a way to check to see if OTHER pages might also not be indexed? What should a small ecom site do to see about getting it listed? SOS in Modesto. Ron
Intermediate & Advanced SEO | | yatesandcojewelers0 -
Factors that affect Google.com vs .ca
Though my company is based in Canada, we have a .com URL, we're hosted on servers in the U.S., and most of our customers are in the U.S. Our marketing efforts are focused on the U.S. Heck, we even drop the "u" in "colour" and "favour"! 🙂 Nonetheless we rank very well in Google.ca, and rather poorly on Google.com. One hypothesis is that we have more backlinks from .ca domains than .com, but I don't believe that to be true. For sure, the highest quality links we have come from .coms like NYTimes.com. Any suggestions on how we can improve the .com rankings, other than keeping on with the link building?
Intermediate & Advanced SEO | | RobM4161