Just read Travis Loncar's YouMoz post and I have a question about Pagination
-
This was a brilliant post.
I have a question about Pagination on sites that are opting to use Google Custom Search. Here is an example of a search results page from one of the sites I work on:
http://www.ccisolutions.com/StoreFront/category/search-return?q=countryman
I notice in the source code of sequential pages that the rel="next" and rel="prev" tags are not used. I also noticed that the URL does not change when clicking on the numbers for the subsequent pages of the search results.
Also, the canonical tag of every subsequent page looks like this:
Are you thinking what I'm thinking? All of our Google Custom Search pages have the same canonical tag....Something's telling me this just can't be good.
Questions:
1. Is this creating a duplicate content issue?
2. If we need to include rel="prev" and rel="next" on Google Custom Search pages as well as make the canonical tag accurate, what is the best way to implement this?
Given that searchers type in such a huge range of search terms, it seems that the canonical tags would have to be somehow dynamically generated.
Or, (best case scenario!) am I completely over-thinking this and it just doesn't matter on dynamically driven search results pages?
Thanks in advance for any comments, help, etc.
-
-
Considering that the larger of the two sites I work on is on a platform from 1996, I might actually be living "back in the day!" lol - Thanks again Jared!
-
This would all depend on what the site was built on, and the flexibility. There's no questions that this can be done. "Back in the day" we had a few sites that had tens of thousands of page due to sorting, and we had everything generated including:
Title, meta d, meta k, breadcrmb, H1 and short description.
Those were the days!!!
-
For the most part, I would choose to use rel=prev/next for pagination, including both pagination with dynamic urls and static URLs. There are some cases (as with this original thread question) where you should use canonical, but as a whole you should use rel=prev/next.
The best way to explain it is:
Rel Prev/Next:
Your site: Hi Google, I have all of these pages that very similar so I'm just letting you know that I only have duplicate content here for usability reasons and am in no way inferring that you should index all of these pages and rank them #1!
Google: Ok great, thanks for letting us know. We'll index the pages we feel are appropriate, but you wont get penalized for duplicate content. We may only index and serve one page, "page 1", or we may index multiple pages. Thanks for letting us know.
Canonical:
Your site: Hi Google, I have all these paginated pages that look like duplicate content, please do not include any of them in your index, and don't penalize me for duplicate content. For the record, the page you should index is Page 1 and no other pages.Any links that point to the paginated pages should be counted towards Page 1*.
Google: Great, no matter what we will not index any pagination and only Page 1.
With rel=next you are simply letting Google know, but not dictating how Google should act on the situation. If fact with ecomm sites, youll find that a lot of timees when you use rel=next, Google will actually index the 'view all' page if you have "view all" as an option around your pagination links
*many articles suggest that link juice is passed to the canonical URL - I'm have not seen any direct evidence of this but is worth a different discussion.
-
Yes, Jared, this is a great answer. I understand completed. It looks like we are ok then with Google Custom Search as it is. Thanks so much for your thoughtful answer. Now, if we can only get our paginated category pages sorted out, we'll be on the right track!
-
Hi Gerd,
Yes, this is a separate issue we are also struggling with on the site. I believe Travis' YouMoz post from yesterday made a pretty good case for using multiple paginated URLs, and he even illustrated how to accomplish this with sorting parameters like "color" and "price"
You raise a very good point about duplicate titles and descriptions potentially being a problem in this scenario.
Does anyone have any ideas about how to handle that? Could the backend be programmed to dynamically create unique titles and descriiptions based on some rules for naming conventions? (assuming you have access to that level of the code of course)
Really interested to know some points of view on this!
Dana
-
I raised a similar question in the following Q&A - http://www.seomoz.org/q/duplicate-title-tags-with-pagination-and-canonical
My concern or question (we have rel=prev/next) would be more towards what the canoncial should be. There seems to be different opinions:
1. Use the current paginated page as the canonical - in our case GWMT reports duplicate titles (I suppose appending a page-number should sort this out)
2. Use the base search URL as the canonical - perhaps not a bad choice if your site's content changes and Google indexes page 50, but over time you only have results for 40 pages (resulting in an empty result page)
I currently only can conclude that having the prev/next implemented is a good thing as it will hint Google in pagination (in addition to setup the URL parameters in GWMT). I do plan to change the canoncial to the base search URL (and not having multiple paginated URLs) and see how this will affect indexing and SERPs.
-
Dana
Great and informative question,
Jared
Great Answer
-
Hi Dana - Let me see if I understand this correctly:
In question 1 you asked if this would be a duplicate content issue. The canonical tag retains the exact same URL regardless of the search parameter (and resulting search results). Therefore, regardless of the search being made, Google and other crawlers will not index page with a search parameter since the canonical references to the original url (http://www.ccisolutions.com/StoreFront/category/search-return). This means that when Google accidentally lands here http://www.ccisolutions.com/StoreFront/category/search-return?q=countryman it sees the canonical tag and understands that it should not index this page as it is only a variation of the core page.
This would of course be a problem if you actually wanted Google to index every query page. Alternate methods could be to disclude the query parameter in WMT or Robots. But the canonical is built in for you so that you dont have to.
In situations like this I also like to add site search to analytics and block the query parameter so no query pages show up as landing pages.
-
I understand exactly what you are saying Jared. However, here's the problem, the canonical tag is exactly the same....for every single subsequent page in a series across the entire site.
No matter what is searched. The canonical tag remains:
Wouldn't that mean that all search results pages, regardless of search term, are viewed as the same page?
I have heard this discussed before come to think of it. In this case, wouldn't it be proper to block all dynamic search results pages from being crawled or indexed by Google via the htaccess file or robots.txt file?
-
Hi Dana -
I think in the case of Google Custom Search, there is no need to worry about duplication. The reason is that although the rel="prev" etc tags are not being used, a blanket solution already exists: the canonical tag. As you mentioned, the canonical tag never changes, regardless of the search - therefore the crawlers only ever see the Custom Search page as a single page regardless of the queries being made. Thus there is no duplicate issue.
-
I use Google custom search on my site and love it. I would say you have some valid concerns. At first it was a bit of a pain because some of the images didn't line up with the products after a few weeks it worked itself out. We had a 47% increase in conversion from using Google custom search, I use an out of the box type web service so I cannot help you with a few of the questions. There is a lot of customization you can do to fix that you described. Bringing our blog and recipe section was the purpose for trying it and the revenue proved it to be a wise decision.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Ajax Pagination in Magento Question
Hi, We just launched our new theme for Magento and my developer stated the pagination uses Ajax. Previously I had the developers set up rel prev/next for all our pages (categories/ecommerce site) that had multiples. He said it's not required with Ajax. Is this correct? Example: https://www.bestpricenutrition.com/whey.html and when you go to Page 2, the URL shows: https://www.bestpricenutrition.com/whey.html? I want to make sure these pages are set up correctly.
Technical SEO | | vetofunk0 -
Question on Google's Site: Search
A client currently has two domains with the same content on each. When I pull up a Cached version of the site, I noticed that it has a Cache of the correct page on it. However, when I do a site: in Google, I am seeing the domain that we don't want Google indexing. Is this a problem? There is no canonical tag and I'm not sure how Google knows to cache the correct website but it does. I'm assuming they have this set in webmaster tools? Any help is much appreciated! Thanks!
Technical SEO | | jeff_46mile0 -
What's Worse - 404 errors or a huge .htaccess file
We have changed our site architecture pretty significantly and now have many fewer pages (albeit with more robust content and focused linking). My question is, what should I do about all the 404 errors (keep in mind, I am only finding these in Bing Webmaster tools, not Moz or GWT)? Is it worse to have all those 404 errors (hundreds), or to have a massive htaccess file for pages that are only getting hits by the Bing crawlbot. Any insight would be great. Thanks
Technical SEO | | CleanEdisonInc0 -
Sitemap question
Hello, In your opinion what is better for a root domain and micro-sites using sub-domains?, to have a single sitemap for the root domain including all links to the sub-domains or to have a separate sitemap for each sub-domain? Thanks Arnold
Technical SEO | | arnoldwender0 -
Duplicate Content - What's the best bad idea?
Hi all, I have 1000s of products where the product description is very technical and extremely hard to rewrite or create an unique one. I'll probably will have to use the contend provided by the brands, which can already be found in dozens of other sites. My options are: Use the Google on/off tags "don't index
Technical SEO | | Carlos-R
" Put the content in an image Are there any other options? We'd always write our own unique copy to go with the technical bit. Cheers0 -
Buying multiple domains: misspells & .net, org, etc. & 301's
Hi, an SEO guy told me to buy up domains like ours X.org, net, biz, etc. & mispellings. this could cost over $100/year. Is is worth it for SEO or is it just covering our @ss if competitors want to get stupid and buy those? I don't forsee competitors doing that. What do you suggest? Does Google actually give us points for those AND if we bought them are we supposed to redirect all of them to our site? Should I be doing this for our SEO clients? Thanks.
Technical SEO | | JCunningham0 -
Blocking URL's with specific parameters from Googlebot
Hi, I've discovered that Googlebot's are voting on products listed on our website and as a result are creating negative ratings by placing votes from 1 to 5 for every product. The voting function is handled using Javascript, as shown below, and the script prevents multiple votes so most products end up with a vote of 1, which translates to "poor". How do I go about using robots.txt to block a URL with specific parameters only? I'm worried that I might end up blocking the whole product listing, which would result in de-listing from Google and the loss of many highly ranked pages. DON'T want to block: http://www.mysite.com/product.php?productid=1234 WANT to block: http://www.mysite.com/product.php?mode=vote&productid=1234&vote=2 Javacript button code: onclick="javascript: document.voteform.submit();" Thanks in advance for any advice given. Regards,
Technical SEO | | aethereal
Asim0 -
Should we use Google's crawl delay setting?
We’ve been noticing a huge uptick in Google’s spidering lately, and along with it a notable worsening of render times. Yesterday, for example, Google spidered our site at a rate of 30:1 (google spider vs. organic traffic.) So in other words, for every organic page request, Google hits the site 30 times. Our render times have lengthened to an avg. of 2 seconds (and up to 2.5 seconds). Before this renewed interest Google has taken in us we were seeing closer to one second average render times, and often half of that. A year ago, the ratio of Spider to Organic was between 6:1 and 10:1. Is requesting a crawl-delay from Googlebot a viable option? Our goal would be only to reduce Googlebot traffic, and hopefully improve render times and organic traffic. Thanks, Trisha
Technical SEO | | lzhao0