Rel=Canonical to Longer Page?
-
We've got a series of articles on the same topic and we consolidated the content and pasted it altogether on a single page. We linked from each individual article to the consolidated page. We put a noindex on the consolidated page.
The problem: Inbound links to individual articles in the series will only count toward the authority of those individual pages, and inbound links to the full article will be worthless.
I am considering removing the noindex from the consolidated article and putting rel=canonicals on each individual post pointing to the consolidated article. That should consolidate the PageRank. But I am concerned about pointing****a rel=canonical to an article that is not an exact duplicate (although it does contain the full text of the original--it's just that it contains quite a bit of additional text).
An alternative would be not to use rel=canonicals, nor to place a noindex on the consolidated article. But then my concern would be duplicate content and unconsolidated PageRank.
Any thoughts?
-
Nice.
-
I am doubting it. Seems like an either/or thing. I'll probably do rel=canonical to the view-all.
Thanks everybody.
-
Now I just want to know if that usage of rel=canonical can coexist with rel=next & rel=prev.
-
"rel=”canonical” can specify the superset of content" -http://googlewebmastercentral.blogspot.com/2011/09/view-all-in-search-results.html
-
Rel=canonical is supposed to point to an identical page.
-
Consolidating for usability - I want to provide access to both formats. I hate the idea of noindexing a page that could be the target of inbound links.
-
Using rel=prev and rel=next would do the trick if it's still cool to point the rel=canonical to the view-all page. Anyone know?
-
Maybe you're better off noindexing the partial articles and linking from them to the main article with a "Read the full article" link, or something like that.
How many of these articles you have relative to the rest of your content could make a difference--a very small percentage probably wouldn't be an issue in the overall health of your site if you just left them as is.
Why are you consolidating?
-
"Should I make the consolidated article a PDF download? Let them all be indexed without canonicals or 301s? I just don't know the best practice here."
First off I would say I would not use a PDF unless you have the photographs and other content to make flow and have a good end-user experience
PDFs in Google search results
Thursday, September 01, 2011 at 7:23 AM
Webmaster level: All
Our mission is to organize the world’s information and make it universally accessible and useful. During this ambitious quest, we sometimes encounter non-HTML files such as PDFs, spreadsheets, and presentations. Our algorithms don’t let different filetypes slow them down; we work hard to extract the relevant content and to index it appropriately for our search results. But how do we actually index these filetypes, and—since they often differ so much from standard HTML—what guidelines apply to these files? What if a webmaster doesn’t want us to index them?
Google first started indexing PDF files in 2001 and currently has hundreds of millions of PDF files indexed. We’ve collected the most often-asked questions about PDF indexing; here are the answers:
Q: Can Google index any type of PDF file?
A: Generally we can index textual content (written in any language) from PDF files that use various kinds of character encodings, provided they’re not password protected or encrypted. If the text is embedded as images, we may process the images with OCR algorithms to extract the text. The general rule of the thumb is that if you can copy and paste the text from a PDF document into a standard text document, we should be able to index that text.Q: What happens with the images in PDF files?
A: Currently the images are not indexed. In order for us to index your images, you should create HTML pages for them. To increase the likelihood of us returning your images in our search results, please read the tips in our Help Center.Q: How are links treated in PDF documents?
A: Generally links in PDF files are treated similarly to links in HTML: they can pass PageRank and other indexing signals, and we may follow them after we have crawled the PDF file. It’s currently not possible to "nofollow" links within a PDF document.Q: How can I prevent my PDF files from appearing in search results; or if they already do, how can I remove them?
A: The simplest way to prevent PDF documents from appearing in search results is to add an X-Robots-Tag: noindex in the HTTP header used to serve the file. If they’re already indexed, they’ll drop out over time if you use the X-Robot-Tag with the noindex directive. For faster removals, you can use the URL removal tool in Google Webmaster Tools.Q: Can PDF files rank highly in the search results?
A: Sure! They’ll generally rank similarly to other webpages. For example, at the time of this post, [mortgage market review], [irs form 2011] or [paracetamol expert report] all return PDF documents that manage to rank highly in our search results, thanks to their content and the way they’re embedded and linked from other webpages.Q: Is it considered duplicate content if I have a copy of my pages in both HTML and PDF?
A: Whenever possible, we recommend serving a single copy of your content. If this isn’t possible, make sure you indicate your preferred version by, for example, including the preferred URL in your Sitemap or by specifying the canonical version in the HTML or in the HTTP headersof the PDF resource. For more tips, read our Help Center article about canonicalization.Q: How can I influence the title shown in search results for my PDF document?
A: We use two main elements to determine the title shown: the title metadata within the file, and the anchor text of links pointing to the PDF file. To give our algorithms a strong signal about the proper title to use, we recommend updating both.If you want to learn more, watch Matt Cutt’s video about PDF files’ optimization for search, and visit our Help Center for information about the content types we’re able to index. If you have feedback or suggestions, please let us know in the Webmaster Help Forum.
As an end-user my personal preference would be to have a regular webpage over a PDF. However a PDF is always going to be indexed course and can be of great value. I personally would go with the individual page versus the extra clicking that it takes to download or view a PDF person does not have the right plug-in install another browser then they cannot appear in their browser and will have to download some people have apprehension about downloading anything at all.For instance this PDF is something that does not rank as well the website counterpart
I hope this is been of help to you sincerely,Thomas
-
I apologize for that confusing post. Using a voice recognition system. So I apologize for any errors
I should've worded my question differently instead of asking I assume that you're using WordPress correct?
I should've asked are you using WordPress or not?
If you were using WordPress in all honesty it would be easier to reference a plug-in for Pagination and that's why I was asking.
What I said below unfortunately it was an error I apologize for that.
"I would of course index your house.
I don't know what this means."
I would of course indexed the webpage is what I meant essentially. I'm sorry for the error
"What exactly should I check using OSE, and what actions should I take in response to what findings?"
Please check your inbound links going to articles that you know are from quality websites if you were to lose those even if they're pointing to partial articles for full articles do you believe that your site would lose page rank?
My reference to open site Explorer was in reply to this you stating the problem with inbound links not counting towards the authority of individual pages and full article inbound links will be worthless.
"The problem: Inbound links to individual articles in the series will only count toward the authority of those individual pages, and inbound links to the full article will be worthless."
Using open site Explorer you can figure out how many inbound links of value are pointing to your articles and they would. OSE gives you the total for your root domain as well as your page and subdomain. Good links regardless of if their pointed at articles or pointed at a secular page would Make a stronger website altogether if they are pointing to any page on your site at all. They would give you stronger domain trust along with stronger page rank.
Please review the articles about new Pagination
However I would strongly recommend handling Pages this way in this article here.
http://searchengineland.com/google-provides-new-options-for-paginated-content-92906
http://googlewebmastercentral.blogspot.com/2011/09/view-all-in-search-results.html
http://googlewebmastercentral.blogspot.com/2011/09/pagination-with-relnext-and-relprev.html
I am very sorry that my first answer was not very helpful to you. I hope this one is of more help and I appreciate you letting me know that I have made an error.
I sincerely hope that this will be of more help to you and answer your question fully.
New Handling of View All Pages
Google has been evolving their detection of a series of component pages and the corresponding view all page. When you have a view all page and paginated URLs with a detectable pattern, Google clusters those together and consolidates the PageRank value and indexing relevance. Basically, all of the paginated URLs are seen as components in a series that rolls up to the view all page. In most cases, Google has found that the best experience for searchers is to rank the view all page in search results. (You can help this process along by using the rel=”canonical” attribute to point all pages to the view all version.)
If You Don’t Want The View All Page To Rank Instead of Paginated URLs
If you don’t want the view all version of your page shown and instead want individual paginated URLs to rank, you can block the view all version with robots.txt or meta noindex. You can also use the all new rel=”next”/rel=”prev” attributes, so read on!
New Pagination Options
If you don’t have a view all page, or you don’t want the view all page to be what appears in search results, you can use the new attributes rel=”next” and rel=”prev” to cluster all of the component pages into a single series. All of the indexing properties for all components in the series are consolidated and the most relevant page in the series will rank for each query. (Yay!)
You can use these attributes for article pagination, product lists, and any other types of pagination your site might have. The first page of the series has only a rel=”next” attribute and the last page of the series has only a rel=”prev” attribute, and all other pages have both. You can still use the rel=”canonical” attribute on all pages in conjunction.
Typically, in this setup, as Google sees all of these component pages as series, the first page of the series will rank, but there may be times when another page is more relevant and will rank instead. In either case, the indexing signals (such as incoming links) are consolidated and shared by the series.
Make sure that the value of rel=”next” and rel=”prev” match the URL (even if it’s non-canonical) as the rel/next values in the series have to match up (you likely will need to dynamically write the values based on the display URL).
There are lots of intricacies to consider here, and I’m working on an in-depth article that runs through everything that came up in the session, so if you have questions, post them here and I’ll add them in!
if you strongly desire your view-all page not to appear in search results: 1) make sure the component pages in the series don’t include rel=”canonical” to the view-all page, and 2) mark the view-all page as “noindex” using any of the standard methods
A few points to mention:
-
The first page only contains rel=”next” and no rel=”prev” markup.
-
Pages two to the second-to-last page should be doubly-linked with both rel=”next” and rel=”prev” markup.
-
The last page only contains markup for rel=”prev”, not rel=”next”.
-
rel=”next” and rel=”prev” values can be either relative or absolute URLs (as allowed by the tag). And, if you include a
<base>
link in your document, relative paths will resolve according to the base URL. -
rel=”next” and rel=”prev” only need to be declared within the section, not within the document .
-
We allow rel=”previous” as a syntactic variant of rel=”prev” links.
-
rel="next" and rel="previous" on the one hand and rel="canonical" on the other constitute independent concepts. Both declarations can be included in the same page. For example, http://www.example.com/article?story=abc&page=2&sessionid=123 may contain:
-
rel=”prev” and rel=”next” act as hints to Google, not absolute directives.
-
When implemented incorrectly, such as omitting an expected rel="prev" or rel="next" designation in the series, we'll continue to index the page(s), and rely on our own heuristics to understand your content.
Sincerely,
Thomas
-
-
Well, the individual articles are entries in a blog while the consolidated article is a separate page altogether. I could 301 the individual article URLs all over to the consolidated article I suppose, but I'd rather not for usability reasons. For the end users I think the ideal really is to have all the individual blog posts as they are, but provide access to the consolidated article.
Should I make the consolidated article a PDF download? Let them all be indexed without canonicals or 301s? I just don't know the best practice here.
-
Since you have consolidated the content and pasted it altogether on a single page who not 301 the old pages to the new consolidated page which has got all the content from the old ones ?
If you can answer NO to all these then I would suggest doing a 301
- Can a user land on the old page and get any information he/she would not have received from the new page?
- Can you think of any reason why the user would want to see the old page ( as opposed to the new one ) ?
-
Expresso, why not just 301 the individual pages to the consolidated article?
-
I assume that you're using WordPress correct?
Nope.
Normally I would say use [rel=canonicals] 100% of the time. For instance if I had a blog post example.com/ABC/
I would obviously use a [rel=canonical]
Same thing if it is example.com/blog/ABC/
use rel=canonicals
To clarify: I am not talking about placing a rel=canonical on a resource located by more than one URL (eg. www.example.com/article-one/ and www.example.com/article-one/?fp=email). I am talking about placing a rel=canonical on each resource in a series, that points to a distinct resource that contains all of the content from all of the resources (eg. www.example.com/article-one/ -> www.example.com/article-all/, www.example.com/article-two/ -> www.example.com/article-all/, etc.).
I would of course index your house.
I don't know what this means.
In all honesty I would check using open site Explorer before you actually change anything. And make sure that your inbound links are not the problem that you're talking about.
What exactly should I check using OSE, and what actions should I take in response to what findings?
-
I assume that you're using WordPress correct?
If you were willing to post the domain I can tell you whether or not you should use rel=canonicals
Normally I would say use the 100% of the time. For instance if I had a blog post example.com/ABC/
I would obviously use a rel=canonicals
Same thing if it is example.com/blog/ABC/
use rel=canonicals
I would of course index your website if it is only showing partial views of your content if you use rel=canonical you will not have to worry about duplicate content issues.
If you're talking about simply changing the post and that will count as a full page I believe you can already do that and you don't have to worry about the page example.com/blog/ taking all the page rank and leaving you with nothing and your articles will rank. However you can simply create new pages instead of new posts in WordPress and that way you would be getting complete inbound link juice to that one secular page.
In all honesty I would check using open site Explorer before you actually change anything. And make sure that your inbound links are not the problem that you're talking about.
http://www.opensiteexplorer.org/
I also recommend using benchwork press hosting from the WPEngine, ZippyKid, Web synthesis, or Pagely they truly do it are worth every cent with the added speed and helpfulness of the WordPress only host.
I hope this is of help sincerely,
Thomas
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is Google ignoring my canonicals?
Hi, We have rel=canonical set up on our ecommerce site but Google is still indexing pages that have rel=canonical. For example, http://www.britishbraces.co.uk/braces/novelty.html?colour=7883&p=3&size=599 http://www.britishbraces.co.uk/braces/novelty.html?p=4&size=599 http://www.britishbraces.co.uk/braces/children.html?colour=7886&mode=list These are all indexed but all have rel=canonical implemented. Can anyone explain why this has happened?
Intermediate & Advanced SEO | | HappyJackJr0 -
Pages with rel "next"/"prev" still crawling as duplicate?
Howdy! I have a site that is crawling as "duplicate content pages" that is really just pagination. The rel next/prev is in place and done correctly but Roger Bot and Google are both showing duplicated content + duplicate page titles & meta's respectively. The only thing I can think of is we have a canonical pointing back at the URL you are on - we do not have a view all option right now and would not feel comfortable recommending it given the speed implications and size of their catalog. Any experience, recommendations here? Something to be worried about? /collections/all?page=15"/>
Intermediate & Advanced SEO | | paul-bold0 -
Is it a problem to use a 301 redirect to a 404 error page, instead of serving directly a 404 page?
We are building URLs dynamically with apache rewrite.
Intermediate & Advanced SEO | | lcourse
When we detect that an URL is matching some valid patterns, we serve a script which then may detect that the combination of parameters in the URL does not exist. If this happens we produce a 301 redirect to another URL which serves a 404 error page, So my doubt is the following: Do I have to worry about not serving directly an 404, but redirecting (301) to a 404 page? Will this lead to the erroneous original URL staying longer in the google index than if I would serve directly a 404? Some context. It is a site with about 200.000 web pages and we have currently 90.000 404 errors reported in webmaster tools (even though only 600 detected last month).0 -
Crawl Issue Found: No rel="canonical" Tags
Given that google have stated that duplicate content is not penalised is this really something that will give sufficient benefits for the time involved?Also, reading some of the articles on moz.com they seem very ambivalent about its use – for example http://moz.com/blog/rel-confused-answers-to-your-rel-canonical-questionsWill any page with a canonical link normally NOT be indexed by google?Thanks.
Intermediate & Advanced SEO | | fdmgroup0 -
Page is noindex
Hi, We set pages with this and i can see in the view source of the page <meta name="robots" content="noindex"/> We had a new page posted in the site and its indexed by Google but now the new post is visible on a page thats shows partial data which we noindexed as above because its duplicate data and search engines dont have to see it But its still crawling Any ideas?
Intermediate & Advanced SEO | | mtthompsons0 -
Canonical and On-Page Report Card
Hello, One quick question about rel canonical. If i use SeoMoz amazing on-page optimization tool i get a grade B if i use www.mydomain.com and my keyword. I get a grade A if i use https://www.mydomain.com and same keyword. I get the grade B coz i don't get the check mark to "Appropriate Use of Rel Canonical" box. Should i use this rel canonical stuff if i am 301 redirecting www. version to https://www. version already. Regards, OÜInigo
Intermediate & Advanced SEO | | InigoOU0 -
Page not appearing in SERPs
I have a regional site that does fairly well for most towns in the area (top 10-20). However, one place that has always done OK and has great content is not anywhere within the first 200. Everything looks OK, canonical link is correct, I can find the page if I search for exact text, there aren't any higher ranking duplicate pages. Any ideas what may have happened and how I can confirm a penalty for example. TIA,
Intermediate & Advanced SEO | | Cornwall
Chris0 -
What are best SEO practices for product pages of unique items when the item is no longer available?
Hello, my company sells used cars though a website. Each vehicle page contains photos and details of the unit, but once the vehicle is sold, all the contents are replaced by a simple text like "this vehicle is not available anymore".
Intermediate & Advanced SEO | | Darioz
Title of the page also change to a generic one.
URL remains the same. I doubt this is the correct way of doing, but I cannot understand what method would be better. The improvement I am considering for pages of no longer available vehicles is this: keep the page alive but with reduced vehicle details, a text like: this vehicles is not available anymore and automatic recommendations for similar items. What do you think? Is this a good practice or do you suggest anything different? Also, should I put a NOINDEX tag on the expired vehicles pages? Thank you in advance for your help.0