Indexing of internal search results: canonicalization or noindex?
-
Hi Mozzers,
First time poster here, enjoying the site and the tools very much.
I'm doing SEO for a fairly big ecommerce brand and an issue regarding internal search results has come up.
www.example.com/electronics/iphone/5s/ gives an overview of the the model-specific listings. For certain models there are also color listings, but these are not incorporated in the URL structure.
Here's what Rand has to say in Inbound Marketing & SEO: Insights From The Moz Blog
Search filters are used to narrow an internal search—it could be price, color, features, etc.
Filters are very common on e-commerce sites that sell a wide variety of products. Search filter
URLs look a lot like search sorts, in many cases:
www.example.com/search.php?category=laptop
www.example.com/search.php?category=laptop?price=1000
The solution here is similar to the preceding one—don’t index the filters. As long as Google
has a clear path to products, indexing every variant usually causes more harm than good.I believe using a noindex tag is meant here.
Let's say you want to point users to an overview of listings for black 5s iphones. The URL is an internal search filter which looks as follows:
www.example.com/electronics/apple/iphone/5s?search=black
Which you wish to link with the anchor text "black iphone 5s".
Correct me if I'm wrong, but if you no-index the black 5s search filters, you lose the equity passed through the link. Whereas if you canonicalize /electronics/apple/iphone/5s you would still leverage the link juice and help you rank for "black iphone 5s". Doesn't it then make more sense to use canonicalization?
-
Hi there,
Just to round this question off, you could canonicalise the query-string URL searching for black iPhones to the iPhone 5s listings page and keep an individual phone's lising at /123456 separate, yes. It's best to keep the canonical tag for truly duplicated or near-duplicated pages, so you would not want to canonicalise an individual product page to a listings page or similar.
-
The tag is good for duplicate content but if /123456 has unique content then you probably don't need the tag on it. I would refrain from trying to implement the tag on ? on larger terms as it will give you a headache.
Some handy tips here- http://moz.com/learn/seo/canonicalization
In Short -
Set up the tag on the filters e.g a page that's the same content but its showing the colour blue then it will feed back the juice to the original but if you've got a page that's not duplicate and has content on it then you could leave it be. Google's pretty clever at working out relationships on pages and duplicate content is not the worse problem for SEO.
Hope that helps!
-
I meant to say that /123456 is an individual listing and /5gs gives an overview of all listings.
Then I could include a canonical tag at /5gs?search=black pointing to /5gs and NOT include a canonical tag at /5gs/123456 because I want the individual listing to rank?
-
Assuming the info is the same content (duplicate) just with a colour etc.
www.example.com/electronics/apple/iphone/5gs/123456
I would put the tag on that page pointing towards:
www.example.com/electronics/apple/iphone/5gs
What the tag is doing is saying the page (123456) is a duplicate of the another page, here is the other page (the link in tag) then Google will put all relevant juice to the original.
The canonical tag is great for duplicate content but it by putting it on a page deeper in the structure it only affects that page not any others. You can sometimes get a bit ahead by trying to canonical pages that don't exists like www.exsample.com?yay
-
Thanks!
I have a follow up question :).
What if there are listings with unique IDs with the following URL structure:
www.example.com/electronics/apple/iphone/5gs/123456
Then, canonicalizing /electronics/apple/iphone/5gs would prevent the listing from ranking.
What is best practice in these cases? Ideally I would like to pass link juice from the ?search filters to the canonical URL but leave the sub-directories as is.
-
Hi there,
Looks like you've gotten to the bottom of it there. The canonical tag is best as you wouldn't loose any link juice but it would get the desired effect of not indexing the filter.
Looks like you've got a handle on it so good luck!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Should I disable the indexing of tags in Wordpress?
Hi, I have a client that is publishing 7 or 8 news articles and posts each month. I am optimising selected posts and I have found that they have been adding a lot of tags (almost like using hashtags) . There are currently 29 posts but already 55 tags, each of which has its own archive page, and all of which are added to the site map to be indexed (https://sykeshome.europe.sykes.com/sitemap_index.xml). I came across an article (https://crunchify.com/better-dont-use-wordpress-tags/) that suggested that tags add no value to SEO ranking, and as a consequence Wordpress tags should not be indexed or included in the sitemap. I haven't been able to find much more reliable information on this topic, so my question is - should I get rid of the tags from this website and make the focus pages, posts and categories (redirecting existing tag pages back to the site home page)? It is a relatively new websites and I am conscious of the fact that category and tag archive pages already substantially outnumber actual content pages (posts and news) - I guess this isn't optimal. I'd appreciate any advice. Thanks wMfojBf
Intermediate & Advanced SEO | | JCN-SBWD0 -
Javascript content not being indexed by Google
I thought Google has gotten better at picking up unique content from javascript. I'm not seeing it with our site. We rate beauty and skincare products using our algorithms. Here is an example of a product -- https://www.skinsafeproducts.com/tide-free-gentle-he-liquid-laundry-detergent-100-fl-oz When you look at the cache page (text) from google none of the core ratings (badges like fragrance free, top free and so forth) are being picked up for ranking. Any idea what we could do to have the rating incorporated in the indexation.
Intermediate & Advanced SEO | | akih0 -
Making Filtered Search Results Pages Crawlable on an eCommerce Site
Hi Moz Community! Most of the category & sub-category pages on one of our client's ecommerce site are actually filtered internal search results pages. They can configure their CMS for these filtered cat/sub-cat pages to have unique meta titles & meta descriptions, but currently they can't apply custom H1s, URLs or breadcrumbs to filtered pages. We're debating whether 2 out of 5 areas for keyword optimization is enough for Google to crawl these pages and rank them for the keywords they are being optimized for, or if we really need three or more areas covered on these pages as well to make them truly crawlable (i.e. custom H1s, URLs and/or breadcrumbs)…what do you think? Thank you for your time & support, community!
Intermediate & Advanced SEO | | accpar0 -
Why is this site not indexed by Google?
Hi all and thanks for your help in advance. I've been asked to take a look at a site, http://www.yourdairygold.ie as it currently does not appear for its brand name, Your Dairygold on Google Ireland even though it's been live for a few months now. I've checked all the usual issues such as robots.txt (doesn't have one) and the robots meta tag (doesn't have them). The even stranger thing is that the site does rank on Yahoo! and Bing. Google Webmaster Tools shows that Googlebot is crawling around 150 pages a day but the total number of pages indexed is zero. It does appear if you carry out a site: search on Google however. The site is very poorly optimised in terms of title tags, unnecessary redirects etc which I'm working on now but I wondered if you guys had any further insights. Thanks again for your help.
Intermediate & Advanced SEO | | iProspect-Ireland0 -
Redirecting index.html to the root
Hi, I was wondering if there is a safe way to consolidate link juice on a single version of a home page. I find incoming links to my site that link to both mysite.com/ and mysite.com/index.html. I've decided to go with mysite.com/ as my main and only URL for the site and now I'd like to transfer all link juice from mysite.com/index.html to mysite.com/
Intermediate & Advanced SEO | | romanbond
When i tried 301 redirect from index.html to the root it created an indefinite loop, of course. I know I can use a RewriteRule.., but will it transfer the juice?? Please help!5 -
Internal or external blog better?
Hello, We are adding content to ourdogsmind(dot)com We're going to have a blog with unique content. Should we use an external blog with links back to our site, or an internal blog. Thanks.
Intermediate & Advanced SEO | | BobGW0 -
Search Engine Pingler
Hello everyone, it's me again 😉 I've just got a Pro membership on SeoMoz and I am full of questions. A few days ago I found very interesting tool called: Search Engine Pingler And description of it was like this: Your website or your page was published a long time, but you can not find it on google. Because google has not index your site. Tool Search engine pingler will assist for you. It will ping the URL of your Page up more than 80 servers of google and other search engines. Inform to the search engine come to index your site. So my question is that tool really helps to increase the indexation of the link by search engine like Google, if not, please explain what is a real purpose of it. Thank you to future guru who can give a right answer 🙂
Intermediate & Advanced SEO | | smokin_ace0 -
Block all search results (dynamic) in robots.txt?
I know that google does not want to index "search result" pages for a lot of reasons (dup content, dynamic urls, blah blah). I recently optimized the entire IA of my sites to have search friendly urls, whcih includes search result pages. So, my search result pages changed from: /search?12345&productblue=true&id789 to /product/search/blue_widgets/womens/large As a result, google started indexing these pages thinking they were static (no opposition from me :)), but i started getting WMT messages saying they are finding a "high number of urls being indexed" on these sites. Should I just block them altogether, or let it work itself out?
Intermediate & Advanced SEO | | rhutchings0