Just read Travis Loncar's YouMoz post and I have a question about Pagination
-
This was a brilliant post.
I have a question about Pagination on sites that are opting to use Google Custom Search. Here is an example of a search results page from one of the sites I work on:
http://www.ccisolutions.com/StoreFront/category/search-return?q=countryman
I notice in the source code of sequential pages that the rel="next" and rel="prev" tags are not used. I also noticed that the URL does not change when clicking on the numbers for the subsequent pages of the search results.
Also, the canonical tag of every subsequent page looks like this:
Are you thinking what I'm thinking? All of our Google Custom Search pages have the same canonical tag....Something's telling me this just can't be good.
Questions:
1. Is this creating a duplicate content issue?
2. If we need to include rel="prev" and rel="next" on Google Custom Search pages as well as make the canonical tag accurate, what is the best way to implement this?
Given that searchers type in such a huge range of search terms, it seems that the canonical tags would have to be somehow dynamically generated.
Or, (best case scenario!) am I completely over-thinking this and it just doesn't matter on dynamically driven search results pages?
Thanks in advance for any comments, help, etc.
-
-
Considering that the larger of the two sites I work on is on a platform from 1996, I might actually be living "back in the day!" lol - Thanks again Jared!
-
This would all depend on what the site was built on, and the flexibility. There's no questions that this can be done. "Back in the day" we had a few sites that had tens of thousands of page due to sorting, and we had everything generated including:
Title, meta d, meta k, breadcrmb, H1 and short description.
Those were the days!!!
-
For the most part, I would choose to use rel=prev/next for pagination, including both pagination with dynamic urls and static URLs. There are some cases (as with this original thread question) where you should use canonical, but as a whole you should use rel=prev/next.
The best way to explain it is:
Rel Prev/Next:
Your site: Hi Google, I have all of these pages that very similar so I'm just letting you know that I only have duplicate content here for usability reasons and am in no way inferring that you should index all of these pages and rank them #1!
Google: Ok great, thanks for letting us know. We'll index the pages we feel are appropriate, but you wont get penalized for duplicate content. We may only index and serve one page, "page 1", or we may index multiple pages. Thanks for letting us know.
Canonical:
Your site: Hi Google, I have all these paginated pages that look like duplicate content, please do not include any of them in your index, and don't penalize me for duplicate content. For the record, the page you should index is Page 1 and no other pages.Any links that point to the paginated pages should be counted towards Page 1*.
Google: Great, no matter what we will not index any pagination and only Page 1.
With rel=next you are simply letting Google know, but not dictating how Google should act on the situation. If fact with ecomm sites, youll find that a lot of timees when you use rel=next, Google will actually index the 'view all' page if you have "view all" as an option around your pagination links
*many articles suggest that link juice is passed to the canonical URL - I'm have not seen any direct evidence of this but is worth a different discussion.
-
Yes, Jared, this is a great answer. I understand completed. It looks like we are ok then with Google Custom Search as it is. Thanks so much for your thoughtful answer. Now, if we can only get our paginated category pages sorted out, we'll be on the right track!
-
Hi Gerd,
Yes, this is a separate issue we are also struggling with on the site. I believe Travis' YouMoz post from yesterday made a pretty good case for using multiple paginated URLs, and he even illustrated how to accomplish this with sorting parameters like "color" and "price"
You raise a very good point about duplicate titles and descriptions potentially being a problem in this scenario.
Does anyone have any ideas about how to handle that? Could the backend be programmed to dynamically create unique titles and descriiptions based on some rules for naming conventions? (assuming you have access to that level of the code of course)
Really interested to know some points of view on this!
Dana
-
I raised a similar question in the following Q&A - http://www.seomoz.org/q/duplicate-title-tags-with-pagination-and-canonical
My concern or question (we have rel=prev/next) would be more towards what the canoncial should be. There seems to be different opinions:
1. Use the current paginated page as the canonical - in our case GWMT reports duplicate titles (I suppose appending a page-number should sort this out)
2. Use the base search URL as the canonical - perhaps not a bad choice if your site's content changes and Google indexes page 50, but over time you only have results for 40 pages (resulting in an empty result page)
I currently only can conclude that having the prev/next implemented is a good thing as it will hint Google in pagination (in addition to setup the URL parameters in GWMT). I do plan to change the canoncial to the base search URL (and not having multiple paginated URLs) and see how this will affect indexing and SERPs.
-
Dana
Great and informative question,
Jared
Great Answer
-
Hi Dana - Let me see if I understand this correctly:
In question 1 you asked if this would be a duplicate content issue. The canonical tag retains the exact same URL regardless of the search parameter (and resulting search results). Therefore, regardless of the search being made, Google and other crawlers will not index page with a search parameter since the canonical references to the original url (http://www.ccisolutions.com/StoreFront/category/search-return). This means that when Google accidentally lands here http://www.ccisolutions.com/StoreFront/category/search-return?q=countryman it sees the canonical tag and understands that it should not index this page as it is only a variation of the core page.
This would of course be a problem if you actually wanted Google to index every query page. Alternate methods could be to disclude the query parameter in WMT or Robots. But the canonical is built in for you so that you dont have to.
In situations like this I also like to add site search to analytics and block the query parameter so no query pages show up as landing pages.
-
I understand exactly what you are saying Jared. However, here's the problem, the canonical tag is exactly the same....for every single subsequent page in a series across the entire site.
No matter what is searched. The canonical tag remains:
Wouldn't that mean that all search results pages, regardless of search term, are viewed as the same page?
I have heard this discussed before come to think of it. In this case, wouldn't it be proper to block all dynamic search results pages from being crawled or indexed by Google via the htaccess file or robots.txt file?
-
Hi Dana -
I think in the case of Google Custom Search, there is no need to worry about duplication. The reason is that although the rel="prev" etc tags are not being used, a blanket solution already exists: the canonical tag. As you mentioned, the canonical tag never changes, regardless of the search - therefore the crawlers only ever see the Custom Search page as a single page regardless of the queries being made. Thus there is no duplicate issue.
-
I use Google custom search on my site and love it. I would say you have some valid concerns. At first it was a bit of a pain because some of the images didn't line up with the products after a few weeks it worked itself out. We had a 47% increase in conversion from using Google custom search, I use an out of the box type web service so I cannot help you with a few of the questions. There is a lot of customization you can do to fix that you described. Bringing our blog and recipe section was the purpose for trying it and the revenue proved it to be a wise decision.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Does anyone know the linking of hashtags on Wix sites does it negatively or postively impact SEO. It is coming up as an error in site crawls 'Pages with 404 errors' Anyone got any experience please?
Does anyone know the linking of hashtags on Wix sites does it negatively or positively impact SEO. It is coming up as an error in site crawls 'Pages with 404 errors' Anyone got any experience please? For example at the bottom of this blog post https://www.poppyandperle.com/post/face-painting-a-global-language the hashtags are linked, but they don't go to a page, they go to search results of all other blogs using that hashtag. Seems a bit of a strange approach to me.
Technical SEO | | Mediaholix0 -
Spam URL'S in search results
We built a new website for a client. When I do 'site:clientswebsite.com' in Google it shows some of the real, recently submitted pages. But it also shows many pages of spam url results, like this 'clientswebsite.com/gockumamaso/22753.htm' - all of which then go to the sites 404 page. They have page titles and meta descriptions in Chinese or Japanese too. Some of the urls are of real pages, and link to the correct page, despite having the same Chinese page titles and descriptions in the SERPS. When I went to remove all the spammy urls in Search Console (it only allowed me to temporarily hide them), a whole load of new ones popped up in the SERPS after a day or two. The site files itself are all fine, with no errors in the server logs. All the usual stuff...robots.txt, sitemap etc seems ok and the proper pages have all been requested for indexing and are slowly appearing. The spammy ones continue though. What is going on and how can I fix it?
Technical SEO | | Digital-Murph0 -
We just can't figure out the right anchor text to use
We have been trying everything we can with anchor text. We have read here that we should try naturalistic language. Our competitors who are above us in Google search results don't do any of this. They only use their names or a single term like "austin web design". Is what we are doing hurting our listings? We don't have any black hat links. Here's what we are doing now. We are going crazy trying to figure this out. We are afraid to do anything in fear it will damage our position. Bob | pallasart web design | 31 | 1,730 |
Technical SEO | | pallasart
| website by pallasart a texas web design company in austin | 15 | 1,526 |
| website by the austin design company pallasart | 14 | 1,525 |
| created by pallasart a web design company in austin texas | 13 | 1,528 |
| created by an austin web design company pallasart | 12 | 1,499 |
| website by pallasart web design an austin web design company | 12 | 1,389 |
| website by pallasart an austin web design company | 11 | 1,463 |
| pallasart austin web design | 9 | 2,717 |
| website created by pallasart a web design company in austin texas | 9 | 1,369 |
| website by pallasart | 8 | 910 |
| austin web design | 5 | 63 |
| pallasart website design austin |0 -
What's the best way to pass link juice to a page on another domain?
I'm working with a non-profit, and their donation form software forces them to host their donation pages on a different domain. I want to attempt to get their donation page to appear in their sitelinks in Google (under the main website's entry), but it seems like the organization's donation forms are at a disadvantage because they're not actually hosted on that site. I know that no matter what I do, there's no way to "force" a sitelink to appear the way I want it, but... I was trying to think if there's a way I can work around this. Do you think 1) creating a url like orgname.org/donate and having that be a 301 redirect to the donation form, and 2) using the /donate redirect all over the site (instead of linking directly to the form) would help? Are there alternatives other folks recommend?
Technical SEO | | clefevre0 -
Google has deindexed 40% of my site because it's having problems crawling it
Hi Last week i got my fifth email saying 'Google can't access your site'. The first one i got in early November. Since then my site has gone from almost 80k pages indexed to less than 45k pages and the number is lowering even though we post daily about 100 new articles (it's a online newspaper). The site i'm talking about is http://www.gazetaexpress.com/ We have to deal with DDoS attacks most of the time, so our server guy has implemented a firewall to protect the site from these attacks. We suspect that it's the firewall that is blocking google bots to crawl and index our site. But then things get more interesting, some parts of the site are being crawled regularly and some others not at all. If the firewall was to stop google bots from crawling the site, why some parts of the site are being crawled with no problems and others aren't? In the screenshot attached to this post you will see how Google Webmasters is reporting these errors. In this link, it says that if 'Error' status happens again you should contact Google Webmaster support because something is preventing Google to fetch the site. I used the Feedback form in Google Webmasters to report this error about two months ago but haven't heard from them. Did i use the wrong form to contact them, if yes how can i reach them and tell about my problem? If you need more details feel free to ask. I will appreciate any help. Thank you in advance C43svbv.png?1
Technical SEO | | Bajram.Kurtishaj1 -
The importance of url's - are they that important?
Hi Guys I'm reading some very contrasting and confusing reviews regarding urls and the impact they have on a sites ability to rank. My client has a number of flooring products, 71 to be exact - categorised under three sub categories 1. Gallery Wood - 2. Prefinshed Wood - 3. Parquet & Reclaimed. All of the 71 products are branded products (names that are completely unrelated to specific keyword search terms. This is having a major impact regarding how we optimise the site. FOR EXAMPLE: A product of the floor called "White Grain" - the "Key Word" we would like to rank this page for is Brown Engineered Flooring. I'm interested to know, should the name of the branded product match the url? What would you change to help this page rank better for the keyword - Brown Engineered Flooring. Title page: White Grain Url: thecompanyname.com/gallery-wood/white-grain (white grain is the name of the product) Key Word: Brown Engineered Flooring **Seo Title: **White Grain, Brown Engineered Flooring by X Meta Description: BLAH BLAH Brown Engineered Flooring BLAH BLAH Any feedback to help get my head around this would be really appreciated. Thank you.
Technical SEO | | GaryVictory0 -
Javascript to manipulate Google's bounce rate and time on site?
I was referred to this "awesome" solution to high bounce rates. It is suppose to "fix" bounce rates and lower them through this simple script. When the bounce rate goes way down then rankings dramatically increase (interesting study but not my question). I don't know javascript but simply adding a script to the footer and watch everything fall into place seems a bit iffy to me. Can someone with experience in JS help me by explaining what this script does? I think it manipulates the reporting it does to GA but I'm not sure. It was supposed to be placed in the footer of the page and then sit back and watch the dollars fly in. 🙂
Technical SEO | | BenRWoodard1 -
Different TLD's same content - duplicate content? - And a problem in foreign googles?
Hi, Operating from the Netherlands with customers troughout Europe we have for some countries the same content. In the netherlands and Belgium Dutch is spoken and in Germany and Switserland German is spoken. For these countries the same content is provided. Does Google see this as duplicate content? Could it be possible that a german customer gets the Swiss website as a search result when googling in the German Google? Thank you for your assistance! kind regards, Dennis Overbeek Dennis@acsi.eu
Technical SEO | | SEO_ACSI0