Techniques to fix eCommerce faceted navigation
-
Hi everyone,
I've read a lot about different techniques to fix duplicate content problems caused by eCommerce faceted navigation (e.g. redundant URL combinations of colors, sizes, etc.). From what I've seen suggested methods include using AJAX or JavaScript to make the links functional for users only and prevent bots from crawling through them.
I was wondering if this technique would work instead?
If we detect that the user is a robot, instead of displaying a link, we simply display its anchor text.
So what would be for a human
COLOR
< li > < a href = red >red < /a > < /li >
< li > < a href = blue>blue < /a > < /li >Would be for a robot
COLOR
< li > red < /li >
< li > blue < /li >Any reason I shouldn't do this?
Thanks!
*** edit
Another reason to fix this is crawl budget since robots can waste their time going through every possible combination of facet. This is also something I'm looking to fix.
-
I share Alan's hesitation - it could look like cloaking, especially if a bot is making the call. If the pages aren't indexed yet, you could just "nofollow" the links - it sends the same signal transparently.
Home Depot is probably pulling it off with the AJAX/JS implementation, which is a bit harder for Google to parse. They also have a massive authority and link profile, so they can always squeak the small stuff by. You might not be so lucky. In general, it's best to stick to the standard practices and not get too tricky.
-
I've been browsing sites looking at what the big players are doing
Homedepot.com seems to be doing exactly this; if you go to
And you click a facet to narrow the result, the page is refreshed via AJAX
If you go to the same page with a Googlebot user agent, even with JavaScript enabled, clicking the checkbox does nothing!
Is this cloaking? Why is this legit?
-
But is it really cloaking? We wouldn't be showing different content. Just disabling links. This article describes a technique that's more akin to cloaking and justifies it because of "intent": http://www.seomoz.org/ugc/dealing-with-faceted-navigation-a-case-study.
The problem with canonical is that the robots will still waste crawl budget going through all the combinations of facets we have. We have hundreds of categories with complex products with 10+ facets with 10+ options each...
-
That would be cloaking, best not do that
A canonical tag would be best, thats what they are for
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Cross Linking two related ecommerce websites
Hi Guys, Hope you'll be able to help me with a technical problem I am facing right now. We are a company right ? We own 2 webistes. Let's say one sells car parts, the other one buys second hand car parts to refurbish them and sell them. (It is not our case, just an example very similar to ours). sellparts.com buyparts.com Both are ecommerce websites, with large catalogues (7000 skus). sellparts sells a lot and is a big actor in its market. buyparts.com doesn't work nad has a really low DA. My new SEO external consultant, which I am not too convinced about, is telling me to cross link the sites on product level using cross-linking extensions. He want have them do-follow. That would mean having hundreds or thousands of links with really similar linking patterns. buy [parts] [model ] [make] sell [parts] [model ] [make] That to me seems a bit too much and I am worried it compromises the sellparts site's SEO. So should i no-follow the links ? Or do it differently ?
Intermediate & Advanced SEO | | Kepass0 -
How to switch from URL based navigation to Ajax, 1000's of URLs gone
Hi everyone, We have thousands of urls generated by numerous products filters on our ecommerce site, eg./category1/category11/brand/color-red/size-xl+xxl/price-cheap/in-stock/. We are thinking of moving these filters to ajax in order to offer a better user experience and get rid of these useless urls. In your opinion, what is the best way to deal with this huge move ? leave the existing URLs respond as before : as they will disappear from our sitemap (they won't be linked anymore), I imagine robots will someday consider them as obsolete ? redirect permanent (301) to the closest existing url mark them as gone (4xx) I'd vote for option 2. Bots will suddenly see thousands of 301, but this is reflecting what is really happening, right ? Do you think this could result in some penalty ? Thank you very much for your help. Jeremy
Intermediate & Advanced SEO | | JeremyICC0 -
Spam Score - how to fix No Contact Info and Low Number of Pages Found
Hi We recently change the layout of our website to a responsive theme in the hope of improve our rankings. We are getting less more traffic but less conversions to sales. Our Spam sore has these flagged No Contact Info - we have the phone number on top of every page and a contact us link at the bottom of every page - Is there something we are missing ? Low Number of Pages Found - we have over 3000 products each of which has a page, plus other info pages on our stite - Whay would this be flagged.RegardsAdrienne
Intermediate & Advanced SEO | | CostumeD0 -
Best to Fix Duplicate Content Issues on Blog If URLs are Set to "No-Index"
Greetings Moz Community: I purchased a SEMrush subscription recently and used it to run a site audit. The audit detected 168 duplicate content issues mostly relating to blog posts tags. I suspect these issues may be due to canonical tags not being set up correctly. My developer claims that since these blog URLs are set to "no-index" these issues do not need to be corrected. My instinct would be to avoid any risk with potential duplicate content. To set up canonicalization correctly. In addition, even if these pages are set to "no-index" they are passing page rank. Further more I don't know why a reputable company like SEMrush would consider these errors if in fact they are not errors. So my question is, do we need to do anything with the error pages if they are already set to "no-index"? Incidentally the site URL is www.nyc-officespace-leader.com. I am attaching a copy of the SEMrush audit. Thanks, Alan BarjWaO SqVXYMy
Intermediate & Advanced SEO | | Kingalan10 -
Should eCommerce Canonicalize to CMS
We have inherited a site that has a Joomla CMS "showroom" front-end and a Magento "store room" for check out etc. Question - As the site's main pages are in the CMS section should we: make all Magento product pages canonical to the main sections/product pages within the CMS (even though there are no duplicate content issues) "No index" the product pages Index but indicate low page value in sitemap Do something else? 🙂 Thanks for any and all input!
Intermediate & Advanced SEO | | TheNorthernOffice790 -
Advice on outranking Amazon and other big names in eCommerce
I have a client that is targeting some product related keywords. They are on page one for them but Amazon, OfficeMax and Staples are ranking in the top 3 spots for this specific product. Before I start targeting completely different words, do you have any advice on how to tackle big name eCommerce sites who are ranking higher than you. Thank you!
Intermediate & Advanced SEO | | TheOceanAgency0 -
Ecommerce best-of-the-web article - big article - navigation tricks
Hello, We're writing our biggest article and trying to make it best-of-the-web. Custom illustrations, comprehensive content, maybe video slideshows How do I help people navigate this big thing? Is there some pretty navigation systems you've seen work? There's a lot of sections and my only idea so far is to use an anchor/id href attribute for each section having a big list of sections at the top of the article.
Intermediate & Advanced SEO | | BobGW0 -
How to fix issues regarding URL parameters?
Today, I was reading help article for URL parameters by Google. http://www.google.com/support/webmasters/bin/answer.py?answer=1235687 I come to know that, Google is giving value to URLs which ave parameters that change or determine the content of a page. There are too many pages in my website with similar value for Name, Price and Number of product. But, I have restricted all pages by Robots.txt with following syntax. URLs:
Intermediate & Advanced SEO | | CommercePundit
http://www.vistastores.com/table-lamps?dir=asc&order=name
http://www.vistastores.com/table-lamps?dir=asc&order=price
http://www.vistastores.com/table-lamps?limit=100 Syntax in Robots.txt
Disallow: /?dir=
Disallow: /?p=
Disallow: /*?limit= Now, I am confuse. Which is best solution to get maximum benefits in SEO?0