Solving pagination issues for e-commerce
-
I would like to ask about a technical SEO issue that may cause duplicate content/crawling issues.
For pagination, how the rel=canonical, rel="prev" rel="next" and noindex tag should be implemented.
Should all three be within the same page source?
Say for example, for one particular category we may have 10 pages of products (product catalogues). So we should noindex page 2 onwards, rel canonical it back to the first page and also rel="prev" and rel="next" each page so Google can understand they contain multiple pages.
If we index these multiple pages it will cause duplicate content issues. But I'm not sure whether all 3 tags need adding.
It's also my understanding that the search results should be noindexed as it does not provide much value as an entry point in search engines.
-
I have found this useful in the past: https://www.ayima.com/guides/conquering-pagination-guide.html
-
Thanks for your advice, I will take a look at the Google webmaster video you've referenced. As we try to rank for specific search terms in our main categories, we put content in there so it can be indexed and it's great for user experience. That's why I was thinking to also implement the rel=canonical tag so the content wasn't duplicated over a series of 10 pages, but if we noindex and use the rel=prev and next tags, that should solve the issue. It's the same for filterable results really, as the content on the page can be duplicated when users choose to filter by specific options, such as size or colour.
-
Hi Joshua,
You will need all 3 of those tags to properly markup your pagination, just not all at the same time.
Page=1 should have a canonical to the base URL (no page=X), and a rel="next" for page 2. Page 2 will have prev tag for the base level URL, and next for page 3. And so on.
Google says they don't index paginated URLs anymore, but I prefer to play it safe and implement these tags anyway.
Regarding this comment: "It's also my understanding that the search results should be noindexed as it does not provide much value as an entry point in search engines." There is some validity to this, but honestly, it's your preference. I lean on the side of preventing indexing of search results. I don't see much value in those pages being indexed, and if you're doing SEO properly, you're already providing solid entry points. Those pages will also use up a lot of your crawl budget, so that's something to consider too. Chances are, there are better sections of your site that you'd prefer bots spend their time on.
-
You shouldn't use rel canonical for pagination - it's main use is to avoid duplicate content issues. It's possible to combine it with rel next/prev but in very specific cases - example can be found here: https://support.google.com/webmasters/answer/1663744?hl=en :
rel="next" and rel="prev" are orthogonal concepts to rel="canonical". You can include both declarations. For example, http://www.example.com/article?story=abc&page=2&sessionid=123 may contain:
=> as you can see the canonical is used to strip the sessionid which could cause duplicate content issues - not to solve the pagination issue
With rel next/previous you indicate to google that the sequence of pages should be considered as one page - which makes sense if you have like 4/5 pages max. If you have a huge number of pages in a pagination this doesn't really make sense. In that case you could just decide to do nothing - or only have the first page indexed - and the other pages have a noindex/follow tag.
Hope this clarifies.
Dirk
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Hreflang implementation issue
We are currently handling search for a global brand www.example.com which has presence in many countries worldwide. To help Google understand that there is an alternate version of the website available in another language, we have used “hreflang” tags. Also, there is a mother website (www.example.com/global) which is given the attribution of “x-default” in the “hreflang” tag. For Malaysia as a geolocation, the mother website is ranking instead of the local website (www.example.com/my) for majority of the products. The code used for “hreflang” tag execution, on a product page, being: These “hreflang” tags are also present in the XML sitemap of the website, mentioning them below: <loc>http://www.example.com/my/product_name</loc> <lastmod>2017-06-20</lastmod> Is this implementation of “hreflang” tags fine? As this implementation is true across all geo-locations, but the mother website is out-ranking me only in the Malaysia market. If the implementation is correct, what could be other reasons for the same ranking issue, as all other SEO elements have been thoroughly verified and they seem fine.
Intermediate & Advanced SEO | | Starcom_Search0 -
Xml sitemap Issue... Xml sitemap generator facilitating only few pages for indexing
Help me I have a website earlier 10,000 WebPages were facilitated in xml sitemap for indexation, but from last few days xml sitemap generator facilitating only 3300 WebPages for indexing. Please help me to resolve the issue. I have checked Google webmaster indexed pages, its showing 8,141. I have tried 2-3 paid tools, but all are facilitating 3300 pages for indexing. I am not getting what is the exact problem, whether the server not allowing or the problem with xml sitemap generator. Please please help me…
Intermediate & Advanced SEO | | udistm0 -
Magento OR OpenCart OR osCommerce OR Zen Cart OR WP e-Commerce OR WooCommerce
Which cms is good for health product website (selling).?
Intermediate & Advanced SEO | | JordanBrown0 -
Https Homepage Redirect & Issue with Googlebot Access
Hi All, I have a question about Google correctly accessing a site that has a 301 redirect to https on the homepage. Here’s an overview of the situation and I’d really appreciate any insight from the community on what the issue might be: Background Info:
Intermediate & Advanced SEO | | G.Anderson
My homepage is set up as a 301 redirect to a https version of the homepage (some users log in so we need the SSL). Only 2 pages on the site are under SSL and the rest of the site is http. We switched to the SSL in July but have not seen any change in our rankings despite efforts increasing backlinks and out put of content. Even though Google has indexed the SSL page of the site, it appears that it is not linking up the SSL page with the rest of the site in its search and tracking. Why do we think this is the case? The Diagnosis: 1) When we do a Google Fetch on our http homepage, it appears that Google is only reading the 301 redirect instructions (as shown below) and is not finding its way over to the SSL page which has all the correct Page Title and meta information. <code>HTTP/1.1 301 Moved Permanently Date: Fri, 08 Nov 2013 17:26:24 GMT Server: Apache/2.2.16 (Debian) Location: https://mysite.com/ Vary: Accept-Encoding Content-Encoding: gzip Content-Length: 242 Keep-Alive: timeout=15, max=100 Connection: Keep-Alive Content-Type: text/html; charset=iso-8859-1 <title>301 Moved Permanently</title> # Moved Permanently The document has moved [here](https://mysite.com/). * * * <address>Apache/2.2.16 (Debian) Server at mysite.com</address></code> 2) When we view a list of external backlinks to our homepage, it appears that the backlinks that have been built after we switched to the SSL homepage have been separated from the backlinks built before the SSL. Even on Open Site, we are only seeing the backlinks that were achieved before we switched to the SSL and not getting to track any backlinks that have been added after the SSL switch. This leads up to believe that the new links are not adding any value to our search rankings. 3) When viewing Google Webmaster, we are receiving no information about our homepage, only all the non-https pages. I added a https account to Google Webmaster and in that version we ONLY receive the information about our homepage (and the other ssl page on the site) What Is The Problem? My concern is that we need to do something specific with our sitemap or with the 301 redirect itself in order for Google to read the whole site as one entity and receive the reporting/backlinks as one site. Again, google is indexing all of our pages but it seems to be doing so in a disjointed way that is breaking down link juice and value being built up by our SSL homepage. Can anybody help? Thank you for any advice input you might be able to offer. -Greg0 -
Number of Links for Internal E-commerce Search Pages (and Anchor Text)
Hello! We have an internal search engine for different email, postal, and phone data products on our website (75,000 product pages... calling all direct marketers!), I've noindexed all our dynamic search pages, but I'm wondering how else I can improve these pages. Should I reduce the amount of links on each page?
Intermediate & Advanced SEO | | Travis-W
Currently there are 20 search results per page. " <variable>Mailing List" has been a pretty good source of traffic for our product pages.
Should I change the anchor text for all the product pages listed to include the added long-tail keyword, or would that be extremely spammy, having the word "Mailing List" 20+ times on my page? We have both static and dynamic search pages - here is one of static ones: http://www.consumerbase.com/direct-marketing-mailing-lists.html
My main problem with adding the long tail KWs to the anchor text is that we still want our static search pages indexed.</variable> Thanks!0 -
Folder or subdomain for new e-commerce addition
Our main content site has 5K pieces of unique content all targeting our market. We are planning to add e-commerce as a source monetizing our audience. Should we place the new commerce platform within a subdirectory or subdomain? The layout we are considering is... shop name: Brand Name Market http://www.brandname.com/market/ http://market.brandname.com I am also considering something like: http://www.brandname.com/market/ aggregating product details and content from http://market.brandname.com/ with rel= back to the subdomain if possible.
Intermediate & Advanced SEO | | ejovi0 -
MOZ crawl report says category pages blocked by meta robots but theyr'e not?
I've just run a SEOMOZ crawl report and it tells me that the category pages on my site such as http://www.top-10-dating-reviews.com/category/online-dating/ are blocked by meta robots and have the meta robots tag noindex,follow. This was the case a couple of days ago as I run wordpress and am using the SEO Category updater plugin. By default it appears it makes categories noindex, follow. Therefore I edited the plugin so that the default was index, follow as I want google to index the category pages so that I can build links to them. When I open the page in a browser and view source the tags show as index, follow which adds up. Why then is the SEOMOZ report telling me they are still noindex,follow? Presumably the crawl is in real time and should pick up the new follow tag or is it perhaps because its using data from an old crawl? As yet these pages aren't indexed by google. Any help is much appreciated! Thanks Sam.
Intermediate & Advanced SEO | | SamCUK0 -
Another E-commerce Canonical Question
Hi guys, Quick question: one of our clients has an e-commerce site with a very poor canonical tag setup and thousands of pages of duplicate content. Let's use this as an example: BRAND > Category > Type > Color
Intermediate & Advanced SEO | | elcrazyhorse
Four separate pages/URLs. The BRAND page lists all products.
The Category page lists all BRAND products for that category.
The Type page lists all BRAND products of a specific type in that category.
The Color page lists all BRAND products of a specific type in that category of a specific color. Anyway, these generate four separate URLs: /BRAND
/BRAND/Category
/BRAND/Category-Type
/BRAND/Category-Type-Color Avoiding duplicate content and product listings, I would appreciate your proposed canonicalization strategy/feedback.0