Canonical url with pagination
-
I would like to find out what is the standard approach for sections of the site with large number of records being displayed using pagination. They don't really contain the same content, but if title tag isn't changed it seem to process it as duplicate content where the parameter in the url indicating the next page is used.
For the time being I've added ' : Page 1' etc. at the end of the title tag for each separate page with the results, but is there a better way of doing it? Should I use the canonical url here pointing to the main page before pagination shows up in the url?
-
Moz crawls paginated pages even if you have added the rel="next" and rel="prev".
-
Does Moz manage crawling through Wordpress paginated posts (with tags rel="next" / "prev") ?
Since I divided long posts in two posts (page 1 and page 2) using "nextpage" feature in Wordpress, Moz shows duplicate title between page 1 and page 2. For example : https://captaincontrat.com/guide/societe-en-cours-de-formation/ and https://captaincontrat.com/guide/societe-en-cours-de-formation/2/
Thanks a lot
-
Thanks.
-
It does, although Google seems to be slightly less fond of it over time. Since I wrote my reply in March, rel=prev/next are actually beginning to be more effective. I've never seen any major issues with NOINDEX'ing pages 2+, though. In many cases, it's just a lot easier to implement.
The big issue this year is that Google sometimes just ignores deindexation signals. So, you really have to try it and see.
I'd also add that I'm talking about search pagination here, not article pagination. Rel=prev/next is a much better choice for article pagination, because the content is unique across pages. Indexing page 11 of search results isn't much of a benefit, in most cases.
-
Anyone use "no-index" and "follow" for page 2 , page 3 etc? Does this work?
-
So, I have to say that I'm actually upset about Google's recent recommendations, because they've presented them as if their simple and definitive, whereas they're actually very complicated to implement and don't always work very well. A couple of problems:
(1) Rel=prev/next is a fairly weak signal. If you're just trying to help the crawlers, it's fine. If you have issues with large-scale duplication (or have been hit with Panda), it's not a good fix, in my experience.
(2) Rel=prev/next isn't honored at all by Bing.
(3) It's actually really tough to code, especially their proposed Rel=prev/next + Rel=canonical solution.
There are a couple of other options:
(a) If you have a "View All" page (or if that's feasible without it being huge), you can rel-canonical to it from all of the paginated pages.
(b) You can META NOINDEX, FOLLOW pages 2+. I find that's a lot easier and usually more effective. Again, it depends on the severity of the problem and scope of the paginated content.
If you're not having problems and can manage the implementation, Rel=prev/next is a decent first step.
I should add that this is assuming you mean internal search results, and not content pagination (like paginated articles). With paginated search, the additional pages usually aren't a good search-user experience (Google visitors don't need to land on Page 11 of 17 of your search results), so I find that proactively managing them is a good thing. It really does depend a lot on the scope and the size of your index, though. This is a very complex issue that tends to get oversimplified.
-
These pages obviously contain different items and each page only shares the same title and meta tags.
Marcin - do you think that if I add the rel attribute that will solve the problem? Will the Moz reports actually pick it and won't mark it as Duplicate Content and Title?
-
Hi Sebastian,
actually, there's a very clean solution which is fully supported by Google - just use rel="next" and rel="prev" in your paginated links to indicate relationships between pages.
Here's a recent discussion of the best practices from Google itself, and here's another comment by Yoast (famous for his Wordpress SEO plugin).
Hope it helps.
-
I think this is going to depend on two things: 1. Your Site Structure and If you want those pages indexed.
Rand Fishkin - recommends for paginated results not to put the canonical tag pointing back to the top page, which I agree.
Site Structure
If the final pages can only be found by going through the paginated structure, you'll definitely want them followed. You'd only want to no-follow to prioritize your crawl rate, but not recommended unless you have multiple formats (see the article above).
Indexed
If the content is unique (usually blog content) and you are getting traffic to those pages from searches then it may be worthwhile to keep them indexed.
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=93710
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Can the design still be considered adaptive if the URL is different?
I was under the impression our site had a mobile dedicated design, but my developers are telling me we have an adaptive design. The mobile site is set up different and has different content and the url is as follows: www.site.com/MobileView/MobileHome.aspx Can it still be considered adaptive if the URL is not the exact same? Hopefully this make sense and I appreciate anyone's input!
Web Design | | AliMac260 -
How would a redesign, content update and URL change affect ranking?
Hi guys, I have a question that I suspect there is no simple true or false answer to, but perhaps someone has done the same thing as we're pondering wether or not to do? We're taking over an existing site that ranks very well on all the important keywords and is obviously very well liked by Google. The site is today hosted on a sub-domain (xxx.domain.com). When taking over, we'll have to redesign the site and recreate most of the content on the site (unique). The site structure, URLs, incoming links etc. will remain exactly the same. Since we are recreating the site, we also have the opportunity to move the site off the sub-domain and on to the main domain (domain.com/xxx - 85/100 Moz rank) and do a 301 Permanent Redirect on all old URLs. Our long-time experience is that content on the main domain, ranks way better than the sub-domain. The big question is wether or not Google will punish us for both changing the content and the location of the site at the same time? Cheers!
Web Design | | mattbs
Matt0 -
301 Redirect Issue for URL with # and !
Hi All, We had a WIX website and now moved to Wordpress. I m having issue while doing redirecting from old URL to new URL. Example: Old Url: http://www.firsttraffic.com.au/#!traffic-management/ccfn New Url: http://www.firsttraffic.com.au/our-services/traffic-management/ I tried different wordpress plugin but nothing works. I m thinking its due to the # . But How can I to redirection for URL like this . Thanks
Web Design | | emarketexperts0 -
Should Blog Category Archive URLs be Set to "No-Index" in Wordpress?
It appears that Google Webmaster Tools is listing about 120 blog archives URLs in Google Index>Index Status that should not be listed. Our site map contains 650 pages, but Google shows 860. Pages like: <colgroup><col width="464"></colgroup>
Web Design | | Kingalan1
| http://www.nyc-officespace-leader.com/blog/category/manhattan-office-space | With Titles Like: <colgroup><col width="454"></colgroup>
| Manhattan Office Space Archives - Metro Manhattan Office Space | Are listed when in the Rogerbot crawl report for the site. How can we remove such pages from Google Webmaster Tools, Index Status? Our site map shows about 650 pages, yet Google show these extra pages. We would prefer that they not be indexed. Note that these pages do not appear when we run a site:www.nyc-officespace-leader.com search. The site has suffered a drop in ranking since May and we feel it prudent to keep Google from indexing useless URLs. Before May 650 pages showed on the Webmaster Tools Index status, and suddenly in early June when we upgraded the site the index grew by about 175 pages. I suspect the 120 blog archives URLs may have something to do with it. How can we get them removed? Can we set them to "No-Index", or should the robot text be used to remove them? Or can some type of removal request be made to Google? My developers have been struggling with this issue since early June. The bloat on the site is about 175 URLs not on the site map. Is there any go to authority on this issue (it is apparently rather complicated) that can provide a definitive answer? Thanks!!
Alan0 -
Existing URL structure and how to handle new pages before migration
Hi there! Currently, our site uses underscores "_" within the url structure. We are moving to Wordpress soon (the site is currently static html) but it will be a couple of months before the migration. Here is an example of the current structure: www.oldsitestructure.com/about_us/success_stories/custom_vinyl_banners When we do change, our url structure will have hyphen's "-" to separate terms, so the preferred new structure will be: www.oldsitestructure.com/about-us/success-stories/custom-vinyl-banners The entire site (with the exception of our Wordpress blog) currently uses the old structure. We have about 10 - 15 pages we will add before our migration, my question is: Should we use the preferred url structure starting NOW or stick with the old one? And set up 301 redirects are part of the migration process? Many thanks!
Web Design | | SEOSponge
Jon0 -
Can you use a base element and mod_rewrite to alleviate the need for absolute URLs?
This is a follow up question to Scott Parsons' question about using absolute versus relative URLs when linking internally. Andy King makes the statement that this can be done and that it saves additional space (which he claims then can improve page speed). Is this a true and accurate statement? Can using a base element and mod-rewrite alleviate the need for absolute URLs? I need to know before going off on a "change all of our relative URLs to absolutes" campaign. Thanks in advance! Dana
Web Design | | danatanseo0 -
URLs with Hashtags - Does Google Index Them?
Hi there, I have a potential issue with a site whereby all pages are dynamically populated using Javascript. Thus, an example of an URL on their site would be www.example.com/#!/category/product. I have read lots of conflicting information on the web - some says Google will ignore everything after the hashtag; other people say that Google will now index everything after the hashtag. Does anybody have any conclusive information about this? Any links to Google or Matt Cutts as confirmation would be brilliant. P.S. I am aware about the potential issue of duplicate content, but I can assure you that has been dealt with. I am only concerned about whether Google will index full URLs that contain hashtags. Thanks all! Mark
Web Design | | markadoi840 -
The use of foreign characters and capital letters in URL's?
Hello all, We have 4 language domains for our website, and a number of our Spanish landing pages are written using Spanish characters - most notably: ñ and ó. We have done our research around the web and realised that many of the top competitors for keywords such as Diseño Web (web design) and Aplicaión iPhone (iphone application) DO NOT use these special chacracters in their URL structure. Here is an example of our URL's EX: http://www.twago.es/expert/Diseño-Web/Diseño-Web However when I simply copy paste a URL that contains a special character it is automatically translated and encoded. EX: http://www.twago.es/expert/Aplicación-iPhone/Aplicación-iPhone (When written out long had it appears: http://www.twago.es/expert/Aplicación-iPhone/Aplicación-iPhone My first question is, seeing how the overwhelming majority of website URL's DO NOT contain special characters (and even for Spanish/German characters these are simply written using the standard English latin alphabet) is there a negative effect on our SEO rankings/efforts because we are using special characters? When we write anchor text for backlinks to these pages we USE the special characteristics in the anchor text (so does most other competitors). Does the anchor text have to exactly I know most webbrowsers can understand the special characters, especially when returning search results to users that either type the special characters within their search query (or not). But we seem to think that if we were doing the right thing, then why does everyone else do it differently? My second question is the same, but focusing on the use of Capital letters in our URL structure. NOTE: When we do a broken link check with some link tools (such as xenu) the URL's that contain the special characters in Spanish are marked as "broken". Is this a related issue? Any help anyone could give us would be greatly appreciated! Thanks, David from twago
Web Design | | wdziedzic0