Canonical Expert question!
-
Hello,
I am looking for some help here with an estate agent property web site. I recently finished the MoZ crawling report and noticed that MoZ sees some pages as duplicate, mainly from pages which list properties as page 1,2,3 etc. Here is an example:
http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=2
http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=3 etc etcNow I know that the best practise says I should set a canonical url to this page:
http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=allbut here is where my problem is.
http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=1 contains good written content (around 750 words) before the listed properties are displayed while the "page=all" page do not have that content, only the properties listed.
Also http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=1 is similar with the originally designed landing page http://www.xxxxxxxxx.com/property-for-rent/london/houses
I would like yoru advise as to what is the best way to can url this and sort the problem. My original thoughts were to can=url to this page http://www.xxxxxxxxx.com/property-for-rent/london/houses instead of the "page=all" version but your opinion will be highly appreciated.
-
Do "/houses" and "/houses?page=1" have exactly the same content? I'd definitely want to see rel=canonical on the "page=1" version - those are just duplicates. Google has expressly said that they don't want you to canonical pages 2, 3, etc. back to page 1. That doesn't mean it never works, just that it's a bit dicey.
As Chris said, rel=prev/next is another option. Theoretically, it would allow all of the results pages to rank, but let Google know they're a series and not count them against you as thin content. In practice, even my enterprise SEO colleagues have mixed feelings. There's just very limited evidence regarding how effective it is. It is low-risk.
The other option is to go a bit more old-school and META NOINDEX anything with "page=", and just let the original version get indexed and rank. This can help prevent any dilution and would also solve your "page=1" issue. The biggest risk here is if that cut off PR flow across your site or if you had links to the paginated results. In most cases, that's unlikely (people don't link to or tweet page 17 of your search results), but it's a case-by-case thing.
Unfortunately, the "best" solution can be very situational, and even Google isn't very clear about it.
-
It would work but the content after that e.g http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=2 would but lost as they would not be indexed. so if there is content on those pages you feel is valuable might want to look int alternatives however is the strongest content is on http://www.xxxxxxxxx.com/property-for-rent/london/houses you will be fine to set that as the tag location.
-
i have but i was hoping to know if this is solved by adding rel=canonical to the original content landing page? http://www.xxxxxxxxx.com/property-for-rent/london/houses
all page have same content but the text content for some reason appears only on http://www.xxxxxxxxx.com/property-for-rent/london/houses page and on http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=1 page
-
Have you considered the paginated tag ? you could also have a page with a view all option and canonical to that and thus get all the content listed. Why wouldn't the view all page have the same content as each page ?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Subdomain question
Hi guys, I have a subdomain on my site that i want to completely remove from the index. I tried already everything to remove it but it is special situation so the only choice i have left is to remove it from Search Console in "Remove URLs" feature. So my question is: if i remove my root subdomain (example: http://subdomain.mydomain.com/) via "Remove URLs" feature in Webmaster Console, will it remove all the URLs coming from that particular domain as well? I also want to make sure that my root domain will stay untouched and be functioning normally. Thank you for advice!
Technical SEO | | odmsoft0 -
Rel = Canonical in Blog Posting
Hello, I keep coming back to rel=canonical issues! I noticed when I "view pagesource" that my drupal blog posting automatically creates link rel="canonical" href="/sample-blog-title" /< pattern (with the > reversed) in the source code. I'm getting a lot of Rel=Canonical warnings and double content warnings from Seomoz so I've been trying to insert link rel="canonical" href="http://www.example.com/blog/my-awesome-blog-post"< but the page won't retain the code for some reason. I'm entering the code in Plain Text, but saving the document as Full HTML. Is there a better piece of code I can put in to demonstrate that the original blog page is the original source? Thanks!
Technical SEO | | OTSEO0 -
Question about Robot.txt
I just started my own e-commerce website and I hosted it to one of the popular e-commerce platform Pinnacle Cart. It has a lot of functions like, page sorting, mobile website, etc. After adjusting the URL parameters in Google webmaster last 3 weeks ago, I still get the same duplicate errors on meta titles and descriptions based from Google Crawl and SEOMOZ crawl. I am not sure if I made a mistake of choosing pinnacle cart because it is not that flexible in terms of editing the core website pages. There is now way to adjust the canonical, to insert robot.txt on every pages etc. however it has a function to submit just one page of robot.txt. and edit the .htcaccess. The website pages is in PHP format. For example this URL: www.mycompany.com has a duplicate title and description with www.mycompany.com/site-map.html (there is no way of editing the title and description of my sitemap) Another error is www.mycompany.com has a duplicate title and description with http://www.mycompany.com/brands?url=brands Is it possible to exclude those website with "url=" and my "sitemap.html" in the robot.txt? or the URL parameters from Google is enough and it just takes a lot of time. Can somebody help me on the format of Robot.txt. Please? thanks
Technical SEO | | paumer800 -
Adding Rel Canonical to multiple pages
Hi, Our CMS generates a lot of duplicate content, (Different versions of every page for 3 different font sizes). There are many other reasons why we should drop this current CMS and go with something else, and we are in the process of doing that. But for now, does anyone know how would I do the following: I've created a spreadsheet that contains the following: Column 1: rel="canonical" tag for URL Column 2: Duplicate Content URL # 1 Column 3: Duplicate Content URL # 2 Column 4: Duplicate Content URL # 3 I want to add the tag from column 1 into the head of every page from column 2,3, and 4. What would be a fast way to do this considering that I have around 1800 rows. Check the screenshot of the builtwith.com result to see more information about the website if that helps. Farris bxySL
Technical SEO | | jdossetti0 -
Google Places phone number question
Hi, A hotel/resort has a main phone number of 1-234-567-8901. This phone number is consistent in over 50 directories. However, they have a spa and restaurant with the same phone number. The front-desk answers the phone and routes the call to either the restaurant or spa. The name of the spa and restaurant are also found in the local listing directories under different DBA's with the same phone number as the Hotel/Resort. For example: ABC Resort - 1-234-567-8901 Spa Cuts - 1-234-567-8901 (same address as ABC Resort) The Spa - 1-234-567-8901 ) same address as ABC Resort) Will this phone number that is used by the 3 separate entities penalize the Google listing placements for the actual Hotel/Resort in Google Places? Thanks everyone!
Technical SEO | | hawkvt10 -
Rel canonical = can it hurt your SEO
I have a site that has been developed to default to the non-www version. However each page has a rel canonical to the non-www version too. Could having this in place on all pages hurt the site in terms of search engines? thanks Steve
Technical SEO | | stevecounsell0 -
Sitemap question
My sitemap includes www.example.com and www.example.com/index.html, they are both the same page, will this have any negative effects, or can I remove the www.example.com/index.html?
Technical SEO | | Aftermath_SEO0 -
Pages not ranking - Linkbuilding Question
It has been about 3 months since we made some new pages, with new, unique copy, but alot of pages (even though they have been indexed) are not ranking in the SERPS I tested it by taking a long snippet of the unique copy form the page and searching for it on Google. Also I checked the ranking using http://arizonawebdevelopment.com/google-page-rank
Technical SEO | | Impact-201555
Which may no be accurate, I know, but would give some indication. The interesting thing was that for the unique copy snippets, sometimes a different page of our site, many times the home page, shows up in the SERP'sSo my questions are: Is there some issue / penalty / sandbox deal with the pages that are not indexed? How can we check that? Or has it just not been enough time? Could there be any duplicate copy issue going on? Shouldn't be, as they are all well written, completely unique copy. How can we check that? Flickr image details - Some of the pages display the same set of images from flickr. The details (filenames, alt info, titles) are getting pulled form flickr and can be seen on the source code. Its a pretty large block of words, which is the same on multiple pages, and uses alot of keywords. Could this be an issue considered duplication or keyword stuffing, causing this. If you think so , we will remove it right away. And then when do we do to improve re-indexing? The reason I started this was because we have a few good opportunities right now for links, and I was wondering what pages we should link to and try to build rankings for. I was thinking about pointing one to /cast-bronze-plaques, but the page is not ranking. The home page, obviously is the oldest page, and ranked the best. The cast bronze plaques page is very new. Would linking to pages that are not ranking well be a good idea? Would it help them to get indexed / ranking? Or would it be better to link to the pages that are already indexed / ranking? If you link to a page that does not seem to be indexed, will it help the domains link profile? Will the link juice still flow through the site0