Specific question about pagination prompted by Adam Audette's Presentation at RKG Summit
-
This question is prompted by something Adam Audette said in this excellent presentation:
http://www.rimmkaufman.com/blog/top-5-seo-conundrums/08062012/
First, I will lay out the issues:
1. All of our paginated pages have the same URL. To view this in action, go here: http://www.ccisolutions.com/StoreFront/category/audio-technica , scroll down to the bottom of the page and click "Next" - look at the URL. The URL is: http://www.ccisolutions.com/StoreFront/IAFDispatcher, and for every page after it, the same URL.
2. All of the paginated pages with non-unique URLs have canonical tags referencing the first page of the paginated series.
3. http://www.ccisolutions.com/StoreFront/IAFDispatcher has been instructed to be neither crawled nor indexed by Google.
Now, on to what Adam said in his presentation: At about minute 24 Adam begins talking about pagination. At about 27:48 in the video, he is discussing the first of three ways to properly deal with pagination issues. He says [I am somewhat paraphrasing]: "Pages 2-N should have self-referencing canonical tags - Pages 2-N should all have their own unique URLs, titles and meta descriptions...The key is, with this is you want deeper pages to get crawled and all the products on there to get crawled too. The problem that we see a lot is, say you have ten pages, each one using rel canonical pointing back to page 1, and when that happens, the products or items on those deep pages don't get get crawled...because the rel canonical tag is sort of like a 301 and basically says 'Okay, this page is actually that page.' All the items and products on this deeper page don't get the love."
Before I get to my question, I'll just throw out there that we are planning to fix the pagination issue by opting for the "View All" method, which Adam suggests as the second of three options in this video, so that fix is coming.
My question is this: It seems based on what Adam said (and our current abysmal state for pagination) that the products on our paginated pages aren't being crawled or indexed. However, our products are all indexed in Google. Is this because we are submitting a sitemap? Even so, are we missing out on internal linking (authority flow) and Google love because Googlebot is finding way more products in our sitemap that what it is seeing on the site? (or missing out in other ways?)
We experience a lot of volatility in our rankings where we rank extremely well for a set of products for a long time, and then disappear. Then something else will rank well for a while, and disappear. I am wondering if this issue is a major contributing factor.
Oh, and did I mention that our sort feature sorts the products and imposes that new order for all subsequent visitors? it works like this: If I go to that same Audio-Technica page, and sort the 125+ resulting products by price, they will sort by price...but not just for me, for anyone who subsequently visits that page...until someone else re-sorts it some other way. So if we merchandise the order to be XYZ, and a visitor comes and sorts it ZYX and then googlebot crawls, google would potentially see entirely different products on the first page of the series than the default order marketing intended to be presented there....sigh.
Additional thoughts, comments, sympathy cards and flowers most welcome.
Thanks all!
-
Hi Dana,
The problem when it comes to passing authority internally is that properly paginated and crawled listing pages can be one of the primary routes via which Google finds and assigns authority to internal pages. Unless those products are linked to elsewhere, they're not going to be found if they cannot be found on a URL like http://www.ccisolutions.com/StoreFront/category/audio-technica?page=2, ?page=3 etc.
The lack of a unique URL with content changed dynamically also means that there never could be a good flow of authority through the site as Google does not have new pages to crawl and new outbound links to index / follow on those pages.
Your diagram is correct - the second option (Page 1 ---authority---> page 2 ----authority---> page 3... ) is what you're looking for with pagination.
-
Thanks so much Jane. I believe that URL is blocked from being crawled by our .htaccess file, although that's something I need to verify with IT. I just know from past discussions that it is blocked from crawling and indexing and it isn't in Google's index.
Would you mind describing, perhaps with a diagram, how this setup is a problem for passing authority internally? I am thinking it breaks the flow kind of like this:
Page 1 of Series -----> passing authority to page 2 --------> authority stops dead in its tracks due to non-unique URL
Instead of looking like this:
Page 1 of Series -----> passing authority to page 2 --------> page 2 passes authority back to home page, page 1 and page 3 of the Series....and so on
Would that be a somewhat accurate description? Thanks so much for responding. It is greatly appreciated!
-
Hi Dana,
Just to be clear, what I'm seeing is that if I visit a page like http://www.ccisolutions.com/StoreFront/category/led-stage-lights with 107 products, and I click "next", I do not receive the same URL with different products, but rather the URL changes to http://www.ccisolutions.com/StoreFront/IAFDispatcher with the next set of results. I think I'm just being blind, but how did you block that URL from crawling and indexing? I can't see a meta tag or a line in the site's robots.txt file.
This set-up definitely needs a modern pagination solution - glad Ryan's post at Ayima helped! This current set-up is definitely very detrimental to passing authority throughout the site to all products effectively.
-
Thank you Jane.
To answer your first question, no, we don't plan on continuing using the current method, however, I am concerned that whatever we decide to change is implemented properly.
To answer the second question, yes, all of our unique product pages return 200 OK status codes. I think the scenario your described is a very plausible scenario and it makes perfect sense to me. Especially you describe something else that I have seen happen, which I didn't even mention in my question. This is, the swapping out of ranking pages. I have seen this many times when one page might rank for a particular term for a while, then drop, and another page will take it's place and may do a lot better or a lot worse, and then in a few months time they will flip flop again.
One thing you didn't speak to was the fact that we have "no crawl, noindex" set up on this URL that is home to all of our paginated series pages. I am wondering what kind of havoc that could be wreaking on our internal linking and authority flow throughout the site? I am thinking it can't possibly be a good thing, no matter how you slice it.
Thank you very much for the link to the Pagination for SEO post. It contained a most excellent flow chart for pagination that I think every SEO should blow up, print out and post on a wall somewhere where developers and IT can see it. I am sharing a small version here:
-
Wow, I had never heard of this method of sorting before! Is this something you plan on keeping?
I would not be surprised if this is somewhat related to why Google sometimes finds and ranks some products, only to drop them for others. If the CMS serves it a different canonical version of the website every time it visits, it will index different products at different times.
The subsequent paginated pages are canonicalised back to the first page, but do you have unique product pages still returning 200 OK? I take it you do, due to the wording of your question but just to clarify: To use Amazon as an example, if a page like this is paginated, is a product page it links to still available even if it is listed on a deeper paginated page? If so, Google won't necessarily drop the product page if it doesn't see it linked to by the paginated listing pages, but it might do if it never sees it again due to pagination and canonicalisation. So you might end up with a situation where Googlebot found the Glastonbury coffee mug in January, keeps ranking it well for a few months, has not seen it again by April and drops it. But a white coffee mug set was found a few times in March, so it ranks that instead.
Have a look at this pagination post from one of my former colleagues as well - it has some comprehensive solutions to ecommerce pagination problems.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
'domain:example.com/' is this line with a '/' at the end of the domain valid in a disavow report file ?
Hi everyone Just out of curiosity, what would happen if in my disavow report I have this line : domain:example.com**/** instead of domain:example.com as recommended by google. I was just wondering if adding a / at the end of a domain would automatically render the line invalid and ignored by Google's disavow backlinks tool. Many thanks for your thoughts
Technical SEO | | LabeliumUSA0 -
Site's IP showing WMT 'Links to My Site'
I have been going through, disavowing spam links in WMT and one of my biggest referral sources is our own IP address. Site: Covers.com
Technical SEO | | evansluke
IP: 208.68.0.72 We have recently fixed a number of 302 redirects, but the number of links actually seems to be increasing. Is this something I should ignore / disavow / fix using a redirect?0 -
Redirecting .edu subdomains to our site or taking the link, what's more valuable?
We have a relationship built through a service we offer to universities to be issued a .edu subdomain that we could redirect to our landing page relevant to that school. The other option is having a link from their website to that same page. My first question is, what would be more valuable? Can you pass domain authority by redirecting a subdomain to a subdirectory in my root domain? Or would simply passing the link equity from a page in their root domain to our page pass enough value? My second question is, if creating a subdomain with a redirect is much more valuable, what is the best process for this? Would we simply have their webmaster create the subdomain for us an have them put a 301 redirect to our page? Is this getting in the greyer hat area? Thanks guys!
Technical SEO | | Dom4410 -
What should I do with a large number of 'pages not found'?
One of my client sites lists millions of products and 100s or 1000s are de-listed from their inventory each month and removed from the site (no longer for sale). What is the best way to handle these pages/URLs from an SEO perspective? There is no place to use a 301. 1. Should we implement 404s for each one and put up with the growing number of 'pages not found' shown in Webmaster Tools? 2. Should we add them to the Robots.txt file? 3. Should we add 'nofollow' into all these pages? Or is there a better solution? Would love some help with this!
Technical SEO | | CuriousCatDigital0 -
Country Specific Domain
Guyz, we are new startups and have one very simple question regarding domain name. Should we use example.com or example.com.au ? Our Goal initially would be to target customer from Australia and gradually go global. So if we opt for .com.au we may have an edge in terms of local SEO in the beginning but lose out in the long run. What is the best way to tackle this? Thanks
Technical SEO | | WayneRooney0 -
Drupal Question
So on our site we have a plugin for our fan gallery. The issue is that I am getting a lot of duplication errors and it's saying the URL is too long and all the errors are coming from the Fan Gallery, which has over 8,000 errors. It seems to be pulling a long form query URL that has over 100 characters. You can't physically see it on the site, but the crawlers can. Anyway I'm trying to figure out a fix for this. One method would be to just stop those pages from being crawled, but I would hate to do that as the fan gallery for us would be a great source of links and content. So I'm wondering if anyone else has had an issue with these types of plugins before where the user can upload a photo or do a video embed and then it submits to the site. If you have a better method please let me know. I usually work on E-comm platforms so my experience with drupal is limited.
Technical SEO | | KateGMaker0 -
Issue with Joomla Site not showing in SERP's
Site: simpsonelectricnc dot com I'm working on a Joomla website for a local business that isn't ranking at all for any relevant keyword - including the business name. The site is only about six months old and has relatively few links. I realize it takes time to compete for even low-volume keywords, but I think something else may be preventing the site from showing up. The site is not blocked by Robots.txt (which includes a valid reference to the sitemap)
Technical SEO | | CGR-Creative
There is no duplicate content issue, the .htaccess is redirecting all non-www traffic to www version
Every page has a unique title and H1 tag.
The URL's are search-engine friendly (not dynamic either)
XML sitemap is live and submitted to Google WM Tools. Google shows that it is indexing about 70% of the submitted URL's. The site has essentially no domain authority (0.02) according to Moz - I'm assuming this is due to lack of links and short life on the web.
Until today, 98% of the pages had identical meta descriptions. Again, I realize six months is not an eternity - but the site will not even show up for "business name + city,state" searches in Google. In fact, the only way I've seen it in organic results is to search for the exact URL. I would greatly appreciate any help.0 -
Handling '?' in URLs.
Adios! (or something), I've noticed in my SEOMoz campaign that I am getting duplicate content warnings for URLs with extensions. For example: /login.php?action=lostpassword /login.php?action=register etc. What is the best way to deal with these type of URLs to avoid duplicate content penelties in search engines? Thanks 🙂
Technical SEO | | craigycraig0