Ecommerce category pages
-
Hi there,
I've been thinking a lot about this lately. I work on a lot of webshops that are made by the same company. I don't like to say this, but not all of their shops perform great SEO-wise.
They use a filtering system which occasionally creates hundreds to thousands of category pages. Basically what happens is this: A client that sells fashion has a site (www.client.com). They have 'main categories' like 'Men' 'Women', 'Kids', 'Sale'.
So when you click on 'men' in the main navigation, you get www.client.com/men/. Then you can filter on brand, subcategory or color. So you get: www.client.com/men/brand. Basically, the url follows the order in which you filter. So you can also get to 'brand' via 'category': www.client.com/shoes/brand
Obviously, this page has the same content as www.client.com/brand/shoes or even /shoes/brand/black and /men/shoes/brand/black if all the brands' shoes happen to be black and mens' shoes.
Currently this is fixed by a dynamic canonical system that canonicalizes the brand/category combinations. So there can be 8000 url's on the site, which canonicalize to about 4000 url's.
I have a gut feeling that this is still not a good situation for SEO, and I also believe that it would be a lot better to have the filtering system default to a defined order, like /gender/category/brand/color so you don't even need to use these excessive amounts of canonicalization. Because, you can canonicalize the whole bunch, but you'd still offer thousands of useless pages for Google to waste its crawl budget on.
Not to mention the time saved when crawling and analysing using Screaming Frog or other audit tools.
Any opinions on this matter?
-
I love this question, Adriaan. It's one that a lot of people have asked over the years and that a lot of people have had to deal with over time especially with ecommerce sites like those you work on.
As you well know, there are multiple ways to handle duplicate content:
- The way you are proposing, which is moving to a static URL structure that always keeps the same order
- A web of canonicals like you seem to have set up (and it sounds like you have it set up correctly)
- The whack-a-mole approach of periodically looking for duplicate content and implementing redirects, which can lead to further issues with internal redirects. This is not a good scalable option.
SEO is all about processes. If you have a canonical process that is working for you and has been scalable (eg you are not manually specifying the URL for each new category created, which is probably done when the merchandising team or feeds update the site), that works to a certain extent.
However, this is like treating a bunch of cuts on your hands with bandaids but not dealing with the fact that a) you only have so much space on your hands and can only apply so many bandaids, and b) that you're still getting cuts on your hands.
I prefer to deal with the root of the issue, which in your case is that you can have multiple URLs targeting the same terms based on the user's (or Googlebot's!) crawl path on your site. I am assuming that you are only putting the canonicals in your XML and HTML sitemaps, by the way?
If I were you, this is how I would tackle your problem:
-
Make sure you are only putting in the canonical URLs to your XML sitemaps. Start here.
-
Do a full crawl of your site and pull all the URLs that are canonicaling elsewhere. Then get your log files and see how much time the search engines are spending on these canonical'd URLs.
-
Also check to see that Google is indeed respecting all of your canonicals! At this scale of canonicals, I'd expect that they are semi-often not respecting them and you are still dealing with duplicate content issues. But again, that's just a hunch I have.
-
Make a decision from there, off of discussions with your engineers/designers/etc about how much work is involved, about if you think it's worthwhile to make the change.
I am **always **a fan of eliminating pages that are canonical'd and not serving a purpose (example: a PPC landing page might be canonical'd and noindexed, and you don't want to remove that page). My suspicion in your case, as well, is that having /brand/mens won't convert any differently from /mens/brand.
At the end of the day, you need to decide how you want your site organized and if your customers (the people buying things on the site) prefer to shop by brand or by gender/sport/whatever. This will help you decide what way to architect your URLs and your site's flow.
Hope that helps!
John
-
Reducing the number of pages that search engines need to crawl is definitely the right way to go, so yeah I would definitely get a uniform URL structure in place if possible. Reduce that crawl budget
-
Thanks for your response Sean. I do know that the use of canonicals is correct here.
My question though, is if it would be better to reduce the amount of actual pages (introduce a uniform URL structure, so to speak) because this would reduce the amount of pages the Google crawler needs to crawl drastically (over 65% on some of my clients webshops). As far as I know, they do crawl every canonicalized url?
-
It does sound like you're adopting a good approach to canonicals. There are a lot of sites out there that do the same approach with non-uniform URL structures such as the one you're using.
Don't suppose you could supply the URL so I can have a look?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Can a duplicate page referencing the original page on another domain in another country using the 'canonical link' still get indexed locally?
Hi I wonder if anyone could help me on a canonical link query/indexing issue. I have given an overview, intended solution and question below. Any advice on this query will be much appreciated. Overview: I have a client who has a .com domain that includes blog content intended for the US market using the correct lang tags. The client also has a .co.uk site without a blog but looking at creating one. As the target keywords and content are relevant across both UK and US markets and not to duplicate work the client has asked would it be worthwhile centralising the blog or provide any other efficient blog site structure recommendations. Suggested solution: As the domain authority (DA) on the .com/.co.uk sites are in the 60+ it would risky moving domains/subdomain at this stage and would be a waste not to utilise the DAs that have built up on both sites. I have suggested they keep both sites and share the same content between them using a content curated WP plugin and using the 'canonical link' to reference the original source (US or UK) - so not to get duplicate content issues. My question: Let's say I'm a potential customer in the UK and i'm searching using a keyword phrase that the content that answers my query is on both the UK and US site although the US content is the original source.
Intermediate & Advanced SEO | | JonRayner
Will the US or UK version blog appear in UK SERPs? My gut is the UK blog will as Google will try and serve me the most appropriate version of the content and as I'm in the UK it will be this version, even though I have identified the US source using the canonical link?2 -
Duplicate Page getting indexed and not the main page!
Main Page: www.domain.com/service
Intermediate & Advanced SEO | | Ishrat-Khan
Duplicate Page: www.domain.com/products-handler.php/?cat=service 1. My page was getting indexed properly in 2015 as: www.domain.com/service
2. Redesigning done in Aug 2016, a new URL pattern surfaced for my pages with parameter "products-handler"
3. One of my product landing pages had got 301-permanent redirected on the "products-handler" page
MAIN PAGE: www.domain.com/service GETTING REDIRECTED TO: www.domain.com/products-handler.php/?cat=service
4. This redirection was appearing until Nov 2016.
5. I took over the website in 2017, the main page was getting indexed and deindexed on and off.
6. This June it suddenly started showing an index of this page "domain.com/products-handler.php/?cat=service"
7. These "products-handler.php" pages were creating sitewide internal duplicacy, hence I blocked them in robots.
8. Then my page (Main Page: www.domain.com/service) got totally off the Google index Q1) What could be the possible reasons for the creation of these pages?
Q2) How can 301 get placed from main to duplicate URL?
Q3) When I have submitted my main URL multiple times in Search Console, why it doesn't get indexed?
Q4) How can I make Google understand that these URLs are not my preferred URLs?
Q5) How can I permanently remove these (products-handler.php) URLs? All the suggestions and discussions are welcome! Thanks in advance! 🙂0 -
One Page Design / Single Product Page
I have been working in a project. Create a framework for multi pages that I have So here is the case
Intermediate & Advanced SEO | | Roman-Delcarmen
Most of them are single page product / one page design wich means that I dont have many pages to optimize. All this sites/ pages follow the rules of a landing page optimization because my main goals is convert as many users as I can. At this point I need to optimize the SEO, the basic stuff such as header, descriptions, tittles ect. But most of my traffic is generated by affiliates, which is good beacuse I dont have to worrie to generate traffic but if the affiliate network banned my product, then I lose all my traffic. Put all my eggs in the same basket is not a good idea. Im not an seo guru so that is the reason Im asking whic strategies and tactics can give me results. All kind of ideas are welcome1 -
How do we decide which pages to index/de-index? Help for a 250k page site
At Siftery (siftery.com) we have about 250k pages, most of them reflected in our sitemap. Though after submitting a sitemap we started seeing an increase in the number of pages Google indexed, in the past few weeks progress has slowed to a crawl at about 80k pages, and in fact has been coming down very marginally. Due to the nature of the site, a lot of the pages on the site likely look very similar to search engines. We've also broken down our sitemap into an index, so we know that most of the indexation problems are coming from a particular type of page (company profiles). Given these facts below, what do you recommend we do? Should we de-index all of the pages that are not being picked up by the Google index (and are therefore likely seen as low quality)? There seems to be a school of thought that de-indexing "thin" pages improves the ranking potential of the indexed pages. We have plans for enriching and differentiating the pages that are being picked up as thin (Moz itself picks them up as 'duplicate' pages even though they're not. Thanks for sharing your thoughts and experiences!
Intermediate & Advanced SEO | | ggiaco-siftery0 -
Base copy on 1 page, then adding a bit more for another page - potential duplicate content. What to do?
Hi all, We're creating a section for a client that is based on road trips - for example, New York to Toronto. We have a 3 day trip, a 5 day trip, a 7 day trip and a 10 day trip. The 3 day trip is the base, and then for the 5 day trip, we add another couple of stops, for the 7 day trip, we add a couple more stops and then for the 10 day trip, there might be two or three times the number of stops of the initial 3 day trip. However, the base content is similar - you start at New York, you finish in Toronto, you likely go through Niagara on all trips. It's not exact duplicate content, but it's similar content. I'm not sure how to look after it? The thoughts we have are:1) Use canonical tags 3,5,7 day trips to the 10 day trip.
Intermediate & Advanced SEO | | digitalhothouse
2) It's not exactly duplicate content, so just go with the content as it is We don't want to get hit by any penalty for duplicate content so just want to work out what you guys think is the best way to go about this. Thanks in advance!0 -
Structure: Should an eCommerce blog have main menu links to each of the store category pages?
Hi, Should my eCommerce site's blog have menu links to the store's category pages? (like in the store itself) The meaning is that every blog post page will have links to category pages that are not related and probably weakens the in-text relevant links. The other option is to have menu links only to the blog category pages and in-article links to the relevant store category pages (maybe add menu button "Go to Store"). Thanks
Intermediate & Advanced SEO | | BeytzNet0 -
Using WP All Import csv import plugin for wordpress to daily update products on large ecommerce site. Category naming and other issues.
We have just got an automated solution working to upload about 4000 products daily to our site. We get a CSV file from the wholesalers server each day and the way they have named products and categories is not ideal. Although most of the products remain the same (don't need to be over written) Some will go out of stock or prices may change etc. Problem is we have no control over the csv file so we need to keep the catagories they have given us. Might be able to create new catgories and have products listed under multiple categories? If anyone has used wp all import or has knoledge in this area please let me know. I have plenty more questions but this should start the ball rolling! Thanks in advance mozzers
Intermediate & Advanced SEO | | weebro0 -
Embedded mobile page?
I have a client who wants a mobile version of their homepage. Normally, I use responsive design to accomplish this for the SEO benefit, but in this case the client wants very different information on the mobile home page than their regular home page. I don't want to go to a dedicated mobile version of the page because they get a fair amount of mobile traffic and so it would probably have a significant negative impact on their SEO to do so. So I was thinking I would add a hidden div to the home page which includes everything they want on the mobile home page and then use CSS to hide the regular content and show the hidden content if someone reaches the page from a smart phone. What do you think about this idea? Would I run afoul of Google's anti-cloaking "rules"? Has anyone done something like this before? Thanks!
Intermediate & Advanced SEO | | farlandlee0