Specific question about pagination prompted by Adam Audette's Presentation at RKG Summit
-
This question is prompted by something Adam Audette said in this excellent presentation:
http://www.rimmkaufman.com/blog/top-5-seo-conundrums/08062012/
First, I will lay out the issues:
1. All of our paginated pages have the same URL. To view this in action, go here: http://www.ccisolutions.com/StoreFront/category/audio-technica , scroll down to the bottom of the page and click "Next" - look at the URL. The URL is: http://www.ccisolutions.com/StoreFront/IAFDispatcher, and for every page after it, the same URL.
2. All of the paginated pages with non-unique URLs have canonical tags referencing the first page of the paginated series.
3. http://www.ccisolutions.com/StoreFront/IAFDispatcher has been instructed to be neither crawled nor indexed by Google.
Now, on to what Adam said in his presentation: At about minute 24 Adam begins talking about pagination. At about 27:48 in the video, he is discussing the first of three ways to properly deal with pagination issues. He says [I am somewhat paraphrasing]: "Pages 2-N should have self-referencing canonical tags - Pages 2-N should all have their own unique URLs, titles and meta descriptions...The key is, with this is you want deeper pages to get crawled and all the products on there to get crawled too. The problem that we see a lot is, say you have ten pages, each one using rel canonical pointing back to page 1, and when that happens, the products or items on those deep pages don't get get crawled...because the rel canonical tag is sort of like a 301 and basically says 'Okay, this page is actually that page.' All the items and products on this deeper page don't get the love."
Before I get to my question, I'll just throw out there that we are planning to fix the pagination issue by opting for the "View All" method, which Adam suggests as the second of three options in this video, so that fix is coming.
My question is this: It seems based on what Adam said (and our current abysmal state for pagination) that the products on our paginated pages aren't being crawled or indexed. However, our products are all indexed in Google. Is this because we are submitting a sitemap? Even so, are we missing out on internal linking (authority flow) and Google love because Googlebot is finding way more products in our sitemap that what it is seeing on the site? (or missing out in other ways?)
We experience a lot of volatility in our rankings where we rank extremely well for a set of products for a long time, and then disappear. Then something else will rank well for a while, and disappear. I am wondering if this issue is a major contributing factor.
Oh, and did I mention that our sort feature sorts the products and imposes that new order for all subsequent visitors? it works like this: If I go to that same Audio-Technica page, and sort the 125+ resulting products by price, they will sort by price...but not just for me, for anyone who subsequently visits that page...until someone else re-sorts it some other way. So if we merchandise the order to be XYZ, and a visitor comes and sorts it ZYX and then googlebot crawls, google would potentially see entirely different products on the first page of the series than the default order marketing intended to be presented there....sigh.
Additional thoughts, comments, sympathy cards and flowers most welcome.
Thanks all!
-
Hi Dana,
The problem when it comes to passing authority internally is that properly paginated and crawled listing pages can be one of the primary routes via which Google finds and assigns authority to internal pages. Unless those products are linked to elsewhere, they're not going to be found if they cannot be found on a URL like http://www.ccisolutions.com/StoreFront/category/audio-technica?page=2, ?page=3 etc.
The lack of a unique URL with content changed dynamically also means that there never could be a good flow of authority through the site as Google does not have new pages to crawl and new outbound links to index / follow on those pages.
Your diagram is correct - the second option (Page 1 ---authority---> page 2 ----authority---> page 3... ) is what you're looking for with pagination.
-
Thanks so much Jane. I believe that URL is blocked from being crawled by our .htaccess file, although that's something I need to verify with IT. I just know from past discussions that it is blocked from crawling and indexing and it isn't in Google's index.
Would you mind describing, perhaps with a diagram, how this setup is a problem for passing authority internally? I am thinking it breaks the flow kind of like this:
Page 1 of Series -----> passing authority to page 2 --------> authority stops dead in its tracks due to non-unique URL
Instead of looking like this:
Page 1 of Series -----> passing authority to page 2 --------> page 2 passes authority back to home page, page 1 and page 3 of the Series....and so on
Would that be a somewhat accurate description? Thanks so much for responding. It is greatly appreciated!
-
Hi Dana,
Just to be clear, what I'm seeing is that if I visit a page like http://www.ccisolutions.com/StoreFront/category/led-stage-lights with 107 products, and I click "next", I do not receive the same URL with different products, but rather the URL changes to http://www.ccisolutions.com/StoreFront/IAFDispatcher with the next set of results. I think I'm just being blind, but how did you block that URL from crawling and indexing? I can't see a meta tag or a line in the site's robots.txt file.
This set-up definitely needs a modern pagination solution - glad Ryan's post at Ayima helped! This current set-up is definitely very detrimental to passing authority throughout the site to all products effectively.
-
Thank you Jane.
To answer your first question, no, we don't plan on continuing using the current method, however, I am concerned that whatever we decide to change is implemented properly.
To answer the second question, yes, all of our unique product pages return 200 OK status codes. I think the scenario your described is a very plausible scenario and it makes perfect sense to me. Especially you describe something else that I have seen happen, which I didn't even mention in my question. This is, the swapping out of ranking pages. I have seen this many times when one page might rank for a particular term for a while, then drop, and another page will take it's place and may do a lot better or a lot worse, and then in a few months time they will flip flop again.
One thing you didn't speak to was the fact that we have "no crawl, noindex" set up on this URL that is home to all of our paginated series pages. I am wondering what kind of havoc that could be wreaking on our internal linking and authority flow throughout the site? I am thinking it can't possibly be a good thing, no matter how you slice it.
Thank you very much for the link to the Pagination for SEO post. It contained a most excellent flow chart for pagination that I think every SEO should blow up, print out and post on a wall somewhere where developers and IT can see it. I am sharing a small version here:
-
Wow, I had never heard of this method of sorting before! Is this something you plan on keeping?
I would not be surprised if this is somewhat related to why Google sometimes finds and ranks some products, only to drop them for others. If the CMS serves it a different canonical version of the website every time it visits, it will index different products at different times.
The subsequent paginated pages are canonicalised back to the first page, but do you have unique product pages still returning 200 OK? I take it you do, due to the wording of your question but just to clarify: To use Amazon as an example, if a page like this is paginated, is a product page it links to still available even if it is listed on a deeper paginated page? If so, Google won't necessarily drop the product page if it doesn't see it linked to by the paginated listing pages, but it might do if it never sees it again due to pagination and canonicalisation. So you might end up with a situation where Googlebot found the Glastonbury coffee mug in January, keeps ranking it well for a few months, has not seen it again by April and drops it. But a white coffee mug set was found a few times in March, so it ranks that instead.
Have a look at this pagination post from one of my former colleagues as well - it has some comprehensive solutions to ecommerce pagination problems.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Microsoft IIS SEO tool claims I have no H1... It's not true.
I have 4300 pages that the tool claims are missing the H1 value but they are there. Here is an example: http://antiquebanknotes.com/rare-currency/first-national-bank-atlanta-illinois-2283.aspx/ Has anyone seen this before?
Technical SEO | | Banknotes0 -
What's Worse - 404 errors or a huge .htaccess file
We have changed our site architecture pretty significantly and now have many fewer pages (albeit with more robust content and focused linking). My question is, what should I do about all the 404 errors (keep in mind, I am only finding these in Bing Webmaster tools, not Moz or GWT)? Is it worse to have all those 404 errors (hundreds), or to have a massive htaccess file for pages that are only getting hits by the Bing crawlbot. Any insight would be great. Thanks
Technical SEO | | CleanEdisonInc0 -
What's another good SEO plugin for WordPress besides Yoast?
I know most people typically use Yoast, we use it on our sites, but we have one site where Yoast is causing a conflict. I wanted to investigate some other options and see what the best solution may be. Anyone else have an effective plugin they like?
Technical SEO | | AdamWormann0 -
How to solve the meta : A description for this result is not available because this site's robots.txt. ?
Hi, I have many URL for commercialization that redirects 301 to an actual page of my companies' site. My URL provider say that the load for those request by bots are too much, they put robots text on the redirection server ! Strange or not? Now I have a this META description on all my URL captains that redirect 301 : A description for this result is not available because this site's robots.txt. If you have the perfect solutions could you share it with me ? Thank You.
Technical SEO | | Vale70 -
Google using descriptions from other websites instead of site's own meta description
In the last month or so, Google has started displaying a description under links to my home page in its search results that doesn't actually come from my site. I have a meta description tag in place and for a very limited set of keywords, that description is displayed, but for the majority of results, it's displaying a description that appears on Alexa.com and a handful of other sites that seem to have copied Alexa's listing, e.g. similarsites.com. The problem is, the description from these other sites isn't particularly descriptive and mentions a service that we no longer provide. So my questions are: Why is Google doing this? Surely that's broken behaviour. How do I fix it?
Technical SEO | | antdesign0 -
'No Follow' and 'Do Follow' links when using WordPress plugins
Hi all I hope someone can help me out with the following question in regards to 'no follow' and 'do follow' links in combination with WordPress plugins. Some plugins that deal with links i.e. link masking or SEO plugins do give you the option to 'not follow' links. Can someone speak from experience that this does actually work?? It's really quite stupid, but only occurred to me that when using the FireFox add on 'NoDoFollow' as well as looking at the SEOmoz link profile of course, 95% of my links are actually marked as FOLLOW, while the opposite should be the case. For example I mark about 90% of outgoing links as no follow within a link masking plugin. Well, why would WordPress plugins give you the option to mark links as no follow in the first place when they do in fact appear as follow for search engines and SEOmoz? Is this a WordPress thing or whatnot? Maybe they are in fact no follow, and the information supplied by SEO tools comes from the basic HTML structure analysis. I don't know... This really got me worried. Hope someone can shed a light. All the best and many thanks for your answers!
Technical SEO | | Hermski0 -
Should we use & or and in our url's?
Example: /Zambia/kasanka-&-bangweulu or /Zambia/kasanka-and-bangweulu which is the better url from the search engines point of view?
Technical SEO | | tribes0 -
URL's for news content
We have made modifications to the URL structure for a particular client who publishes news articles in various niche industries. In line with SEO best practice we removed the article ID from the URL - an example is below: http://www.website.com/news/123/news-article-title
Technical SEO | | mccormackmorrison
http://www.website.com/news/read/news-article-title Since this has been done we have noticed a decline in traffic volumes (we have not as yet assessed the impact on number of pages indexed). Google have suggested that we need to include unique numerical IDs in the URL somewhere to aid spidering. Firstly, is this policy for news submissions? Secondly (if the previous answer is yes), is this to overcome the obvious issue with the velocity and trend based nature of news submissions resulting in false duplicate URL/ title tag violations? Thirdly, do you have any advice on the way to go? Thanks P.S. One final one (you can count this as two question credits if required), is it possible to check the volume of pages indexed at various points in the past i.e. if you think that the number of pages being indexed may have declined, is there any way of confirming this after the event? Thanks again! Neil0