Pagination & duplicate meta
-
Hi
I have a few pages flagged for duplicate meta e.g.:
http://www.key.co.uk/en/key/workbenches?page=2
http://www.key.co.uk/en/key/workbenchesI can;t see anything wrong with the pagination & other pages have the same code, but aren't flagged for duplicate:
http://www.key.co.uk/en/key/coshh-cabinets
http://www.key.co.uk/en/key/coshh-cabinets?page=2
I can't see to find the issue - any ideas?
Becky
-
Regarding the links which point to pages, but include the hash. If Google is only seeing this page http://www.key.co.uk/en/key/workbenches
Will it be seeing these as pages which have duplicate content?
-
No problem thank you
-
I could write out how to implements rel next prev but it would be better to look at these articles
https://webmasters.googleblog.com/2011/09/pagination-with-relnext-and-relprev.html
https://moz.com/ugc/seo-guide-to-google-webmaster-recommendations-for-pagination
-
Hi,
Yes there is javascript to sort the results on those pages.
Is the solution to have these URLs page=2 etc, correctly linked from the page number?
Then ensure rel/prev are used correctly?
I'm also concerned about the content we have at the bottom of the products being shown as duplicate.
-
Hi
Thank you for this. One thing I am confused about is, if Google doesn't crawl those paginated pages, why will it pick up the meta as duplicate?
Thank you for highlighting the links - I hadn't noticed this before.
Where should the rel next prev be coded?
Thanks for your feedback
-
I get how hashes work.
Crawlers do see the page=2, page=3, etc. URLs because the right/left navigation buttons to the side of the numbers link to them. I just proved this by crawling the site in Screaming Frog and doing a search for page=, they're all found.
Becky, there's something larger at play here, potentially with your CMS configuration. It looks like the navigation for paginated sections is messed up. Mouse-over the links and look at the URL in the lower left of your browser, and then click the link and look at your URL bar. The results are very different from what you see on mouse-over. I'd recommend your first step is to talk to your developers and see if they can fix this issue. As VivaCa mentioned, you could be getting false alarms on duplicates here from Moz, so you might be clear with the canonical and prev/next fix - Screaming Frog finds all of those tags properly.
-
I think you guys are missing the point. Anything after the hashtag is ignored. As far as the crawler is concerned, all the links to page 2,3,4,5 are all the same URL - that is why the crawler does not see the other pages.
There is no issue with canonical or how it interacts with the rel next prev. My point on the canonical was simply for illustrative purposes and looks to be implemented correctly.
Separate from the canonical the rel next prevs are setup incorrectly and that needs to be fixed once the issue with how the paginated pages are linked to using the URL with the hashtag parameters.
-
We have the exact same issue, and I found this reply from Dr. Pete helpful regarding this (assuming that what he says is still true): https://moz.com/community/q/pagination-issues-on-e-commerce-site-duplicate-page-title-and-content-on-moz-crawl
His reply:
Unfortunately, Moz Analytics/PRO don't process rel=prev/next properly at this time, so we may give false alarms on those pages, even if the tags are properly implemented.
It can be tricky, but Google recommends a combination of rel=canonical and rel=prev/next. Use the canonical tag to keep sorts from getting indexed, and then use rel=prev/next for the pagination itself. Your 3rd example (page=2...) should rel=prev/next to the URLs before and after it but then canonical to the page=2 variation with no sort parameter. It can get complicated fast, unfortunately, but typically rel=canonical can be implemented in the template. So, once you've got it figured out, it'll work for the entire site.
-
As far as I am aware, there is nothing wrong with using both canonicals and pagination on the same page. Google says this as well here: https://support.google.com/webmasters/answer/1663744?hl=en
We have pagination and canonicals set up as suggested in the Google article and also have some issues with Moz saying we have duplicate content, which the pagination should "fix" as far as I understand it.
From the article:
rel="next"
andrel="prev"
are orthogonal concepts torel="canonical"
. You can include both declarations. For example, http://www.example.com/article?story=abc&page=2&sessionid=123 may contain: -
View source on both pages.
http://www.computerhope.com/issues/ch000746.htm
Or use the handy Moz bar to view the descriptions
Both your title and meta are exactly the same - aka they are duplicates
view-source:http://www.key.co.uk/en/key/workbenches
<title>Workbenches & Work Stations from Key</title>view-source:http://www.key.co.uk/en/key/workbenches?page=2
<title>Workbenches & Work Stations from Key</title>You can remedy this by simply adding "- Page #" at the end of your title and description, where # is whatever page in the pagination you are at.
The reason why the other pages in your pagination are not showing up with the duplicate issue is that you are hiding them from Google.
When I am on Page 2 and I click on the buttons for page 3,4,5 etc - here are the links that are shown
Page 3 http://www.key.co.uk/en/key/workbenches?page=2#productBeginIndex:60&orderBy:5&pageView:list&
Page 4 http://www.key.co.uk/en/key/workbenches?page=2#productBeginIndex:90&orderBy:5&pageView:list&
Page 5 http://www.key.co.uk/en/key/workbenches?page=2#productBeginIndex:120&orderBy:5&pageView:list&
These are the links that people can click on to navigate at the bottom of the page. Everything behind the hash is ignored by Google. It is a clever way to hide parameters, but when Google looks at this it is just seeing links to the exact same page. Likewise, on that page you have a canonical link to page 2, so even if Google could see the parameters you are giving it a directive that tells Google that Page 2 is the only page that exists.
I can see that you are using rel next prev to designate Page 3 as Page 3 http://www.key.co.uk/en/key/workbenches?page=3 etc, but you are not coding the rel next prev properly by putting it up in the header with the meta tags.
In summary
- You have duplicate title and meta tags for all your paginated pages
- You are not linking to your paginated pages properly within the user navigation
- You are incorrectly using rel next prev
-
Hi,
I can't explain why Moz throws a duplicate for one and not the other, that's odd. I did look at the source code for both of the paginated URLs you posted, and it looks like rel=prev/next is mostly right, but a couple suggestions:
- Remove the self-referring canonical tags - On this URL (http://www.key.co.uk/en/key/coshh-cabinets?page=2) you've got a canonical that points to itself, that's in conflict with the rel=prev/next tags. Rel=prev/next should be used in place of canonical tags, not in conjunction with.
- The one exception to my point about canonicals above: on page=1 of your pagination, canonicalize that to the root. Example, http://www.key.co.uk/en/key/coshh-cabinets?page=1 should canonicalize to http://www.key.co.uk/en/key/coshh-cabinets, since those are identical in actual displayed content.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate content across domains?
Does anyone have suggestions for managing duplicate product/solution website content across domains? (specifically parent/child company domains) Is it advisable to do this? Will it hurt either domain? Any best practices when going down this path?
Intermediate & Advanced SEO | | pilgrimquality0 -
Related products & SEO
My company has a comprehensive set of historical images and text - hosted separately on a free museum site - it's currently displayed on our main site as an iframe. I realize the iframe brings no SEO juice to the site - but we are updating our site - and thinking of bringing the images and text to our site. I'm wondering if this could help or hurt us - the historical information is about "boat widgets" and we sell "car widgets" - could a lot of information about "boat widgets" dilute our "car widgets" seo ?
Intermediate & Advanced SEO | | ThomasErb0 -
Opinion on Duplicate Content Scenario
So there are 2 pest control companies owned by the same person - Sovereign and Southern. (The two companies serve different markets) They have two different website URLs, but the website code is actually all the same....the code is hosted in one place....it just uses an if/else structure with dynamic php which determines whether the user sees the Sovereign site or the Southern site....know what I am saying? Here are the two sites: www.sovereignpestcontrol.com and www.southernpestcontrol.com. This is a duplicate content SEO nightmare, right?
Intermediate & Advanced SEO | | MeridianGroup0 -
Duplicate Content Question
My understanding of duplicate content is that if two pages are identical, Google selects one for it's results... I have a client that is literally sharing content real-time with a partner...the page content is identical for both sites, and if you update one page, teh otehr is updated automatically. Obviously this is a clear cut case for canonical link tags, but I'm cuious about something: Both sites seem to show up in search results but for different keywords...I would think one domain would simply win out over the other, but Google seems to show both sites in results. Any idea why? Also, could this duplicate content issue be hurting visibility for both sites? In other words, can I expect a boost in rankings with the canonical tags in place? Or will rankings remain the same?
Intermediate & Advanced SEO | | AmyLB0 -
Virtual Domains and Duplicate Content
So I work for an organization that uses virtual domains. Basically, we have all our sites on one domain and then these sites can also be shown at a different URL. Example: sub.agencysite.com/store sub.brandsite.com/store Now the problem comes up often when we move the site to a brand's URL versus hosting the site on our URL, we end up with duplicate content. Now for god knows what damn reason, I currently cannot get my dev team to implement 301's but they will implement 302's. (Dont ask) I also am left with not being able to change the robots.txt file for our site. They say if we allowed people to go in a change this stuff it would be too messy and somebody would accidentally block a site that was not supposed to be blocked on our domain. (We are apparently incapable toddlers) Now I have an old site, sub.agencysite.com/store ranking for my terms while the new site is not showing up. So I am left with this question: If I want to get the new site ranking what is the best methodology? I am thinking of doing a 1:1 mapping of all pages and set up 302 redirects from the old to the new and then making the canonical tags on the old to reflect the new. My only thing here is how will Google actually view this setup? I mean on one hand I am saying
Intermediate & Advanced SEO | | DRSearchEngOpt
"Hey, Googs, this is just a temp thing." and on the other I am saying "Hey, Googs, give all the weight to this page, got it? Graci!" So with my limited abilities, can anybody provide me a best case scenario?0 -
Penalised for duplicate content, time to fix?
Ok, I accept this one is my fault but wondering on time scales to fix... I have a website and I put an affiliate store on it, using merchant datafeeds in a bid to get revenue from the site. This was all good, however, I forgot to put noindex on the datafeed/duplicate content pages and over a period of a couple of weeks the traffic to the site died. I have since nofollowed or removed the products but some 3 months later my site still will not rank for the keywords it was ranking for previously. It will not even rank if I type in the sites' name (bright tights). I have searched for the name using bright tights, "bright tights" and brighttights but none of them return the site anywhere. I am guessing that I have been hit with a drop x place penalty by Google for the duplicate content. What is the easiest way around this? I have no warning about bad links or the such. Is it worth battling on trying to get the domain back or should I write off the domain, buy a new one and start again but minus the duplicate content? The goal of having the duplicate content store on the site was to be able to rank the category pages in the store which had unique content on so there were no problems with that which I could foresee. Like Amazon et al, the categories would have lists of products (amongst other content) and you would click through to the individual product description - the duplicate page. Thanks for reading
Intermediate & Advanced SEO | | Grumpy_Carl0 -
Duplicate content that looks unique
OK, bit of an odd one. The SEOmoz crawler has flagged the following pages up as duplicate content. Does anyone have any idea what's going on? http://www.gear-zone.co.uk/blog/november-2011/gear$9zone-guide-to-winter-insulation http://www.gear-zone.co.uk/blog/september-2011/win-a-the-north-face-nuptse-2-jacket-with-gear-zone http://www.gear-zone.co.uk/blog/july-2011/telephone-issues-$9-2nd-july-2011 http://www.gear-zone.co.uk/blog/september-2011/gear$9zone-guide-to-nordic-walking-poles http://www.gear-zone.co.uk/blog/september-2011/win-a-the-north-face-nuptse-2-jacket-with-gear-zone https://www.google.com/webmasters/tools/googlebot-fetch?hl=en&siteUrl=http://www.gear-zone.co.uk/
Intermediate & Advanced SEO | | neooptic0 -
Subdomains - duplicate content - robots.txt
Our corporate site provides MLS data to users, with the end goal of generating leads. Each registered lead is assigned to an agent, essentially in a round robin fashion. However we also give each agent a domain of their choosing that points to our corporate website. The domain can be whatever they want, but upon loading it is immediately directed to a subdomain. For example, www.agentsmith.com would be redirected to agentsmith.corporatedomain.com. Finally, any leads generated from agentsmith.easystreetrealty-indy.com are always assigned to Agent Smith instead of the agent pool (by parsing the current host name). In order to avoid being penalized for duplicate content, any page that is viewed on one of the agent subdomains always has a canonical link pointing to the corporate host name (www.corporatedomain.com). The only content difference between our corporate site and an agent subdomain is the phone number and contact email address where applicable. Two questions: Can/should we use robots.txt or robot meta tags to tell crawlers to ignore these subdomains, but obviously not the corporate domain? If question 1 is yes, would it be better for SEO to do that, or leave it how it is?
Intermediate & Advanced SEO | | EasyStreet0