Rel="prev" and rel="next" implementation
-
Hi there
since I've started using semoz I have a problem with duplicate content so I have implemented on all the pages with pagination rel="prev" and rel="next" in order to reduce the number of errors but i do something wrong and now I can't figure out what it is.
the main page url is : alegesanatos.ro/ingrediente/
and for the other pages :
alegesanatos.ro/ingrediente/p2/ - for page 2
alegesanatos.ro/ingrediente/p3/ - for page 3 and so on.
We've implemented rel="prev" and rel="next" according to google webmaster guidelines without adding canonical tag or base link in the header section and we still get duplicate meta title error messages for this pages.
Do you think there is a problem because we create another url for each page instead of adding parameters (?page=2 or ?page=3 ) to the main url
alegesanatos.ro/ingrediente?page=2
thanks
-
Technically, rel=prev/next doesn't de-duplicate the way the canonical tag does, but it should solve any problems for Google. I don't believe we currently consider rel=prev/next when determining duplicate titles. Klarke is right - you could just give those pages semi-unique titles. We're not handling rel=prev/next as well as we could be (it turns out to be a tricky tag to parse well).
Looking at your pages, your implementation appears to be correct. My gut reaction is that your probably ok here. You're doing what Google claims they want (at least what they want this week).
-
Adding rel=next/rel=prev, will not fix issues with Duplicate Titles. You need to edit your templates so that paginated pages actually have unique titles. For example
Ingrediente - Page 1 of 5
Ingrediente - Page 2 of 5
Ingrediente - Page 3 of 5
-
Using rel=prev and rel=next will help to avoid duplicate content issues and is entirely separate from your meta tags. If you wish to avoid duplicate meta title errors then you will need to add 'page x' to your titles, similar to what Matt Cutts has done with his blog.
Hope that helps,
Adam.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is it good practice to use "SAVE $1000's" in SEO titles and Meta Descriptions?
Our company sells a product system that will permanently waterproof almost anything. We market it as a DIY system. I am working on SEO titles and descriptions. This topic came up for discussion, if using "SAVE $1000's.." would help or hurt. We are trying to create an effective call to action, but we are wondering if search engines see it as click bait. Can you
Intermediate & Advanced SEO | | tyler.louth0 -
Website was hacked and is clean now. What to do next for Google?
Hai All, Website was hacked and is clean now. What to do next? Send a Reconsideration Request to Google through Google search console. How detailed should this request be? As detailed as possible? And how much time will it usually take before Google responds? Should I remove all infected URL’s that Google has indexed by using Google search console, remove URL’s? Or is this not necessary? (All the infected URL’s will give a 404 statuscode now) Regards, Maurice
Intermediate & Advanced SEO | | mlehr0 -
Rel=prev/next and canonical tags on paginated pages?
Hi there, I'm using rel="prev" and rel="next" on paginated category pages. On 1st page I'm also setting a canonical tag, since that page happens to get hits to an URL with parameters. The site also uses mobile version of pages on a subdomain. Here's what markup the 1st desktop page has: Here's what markup the 2nd desktop page has: Here's what markup the 1st MOBILE page has: Here's what markup the 2nd MOBILE page has: Questions: 1. On desktop pages starting from page 2 to page X, if these pages get traffic to their versions with parameters, will I'll have duplicate issues or the canonical tag on 1st page makes me safe? 2. Should I use canonical tags on mobile pages starting from page 2 to page X? Are there any better solutions of avoiding duplicate content issues?
Intermediate & Advanced SEO | | poiseo1 -
Received "Googlebot found an extremely high number of URLs on your site:" but most of the example URLs are noindexed.
An example URL can be found here: http://symptom.healthline.com/symptomsearch?addterm=Neck%20pain&addterm=Face&addterm=Fatigue&addterm=Shortness%20Of%20Breath A couple of questions: Why is Google reporting an issue with these URLs if they are marked as noindex? What is the best way to fix the issue? Thanks in advance.
Intermediate & Advanced SEO | | nicole.healthline0 -
Is it ok to use both 301 redirect and rel="canonical' at the same time?
Hi everyone, I'm sorry if this has been asked before. I just wasn't able to find a response in previous questions. To fix the problems in our website regarding duplication I have the possibility to set up 301's and, at the same time, modify our CMS so that it automatically sets a rel="canonical" tag for every page that is generated. Would it be a problem to have both methods set up? Is it a problem to have a on a page that is redirecting to another one? Is it advisable to have a rel="canonical" tag on every single page? Thanks for reading!
Intermediate & Advanced SEO | | SDLOnlineChannel0 -
Fetch as GoogleBot "Unreachable Page"
Hi, We are suddenly having an error "Unreachable Page" when any page of our site is accessed as Googlebot from webmaster tools. There are no DNS errors shown in "Crawl Errors". We have two web servers named web1 and web2 which are controlled by a software load balancer HAProxy. The same network configuration has been working for over a year now and never had any GoogleBot errors before 21st of this month. We tried to check if there could be any error in sitemap, .htaccess or robots.txt by excluding the loadbalancer and pointing DNS to web1 and web2 directly and googlebot was able to access the pages properly and there was no error. But when loadbalancer was made active again by pointing the DNS to it, the "unreachable page" started appearing again. This very same configuration has been working properly for over a year till 21st of this month. Website is properly accessible from browser and there are no DNS errors either as shown by "Crawl Errors". Can you guide me about how to diagnose the issue. I've tried all sorts of combinations, even removed the firewall but no success. Is there any way to get more details about error instead of just "Unreachable Page" error ? Regards, shaz
Intermediate & Advanced SEO | | shaz_lhr0 -
Tumblr the next Squidoo?
What does SEO Moz think of all this SEO abuse that is happening on tumblr lately? I mean, it doesn't take much time to look around tumbler and realize pretty much every page looks like this: http://staff.tumblr.com/post/454906416/blackberry-app Below the post check out the "reblogged". With the correct settings in your tumblr profile, all you gotta do is click reblog on a couple tumblr pages and you get instant backlinks to your website from tumblr pages. Some examples from this page: aticards reblogged this from staff mymuffincorner reblogged this from staff contextual-links reblogged this from staff It's actually a followed link if you reblog. if you just 'like' you end up with a nofollowed link. Questions: Why is tumblr so blind to allow this to happen Surely google must be aware of this and devaluing links that come from tumblr posts? Thoughts? edit: 50 views and no responses. maybe my question isnt clear enough?
Intermediate & Advanced SEO | | adriandg0 -
How Can I know that a link placed is not lableld "No Follow"?
If someone wants to trade links, how can I be sure the link is followed?
Intermediate & Advanced SEO | | SEObleu.com0