Reinforcing Rel Canonical? (Fixing Duplicate Content)
-
Hi Mozzers,
We're having trouble with duplicate content between two sites, so we're looking to add some oomph to the rel canonical link elements we put on one of our sites pointing towards the other to help speed up the process and give Google a bigger hint.
Would adding a hyperlink on the "copying" website pointing towards the "original" website speed this process up?
Would we get in trouble if added about 80,000 links (1 on each product page) with a link to the matching product on the other site? For example, we could use text like "Buy XY product on Other Brand Name and receive 10% off!"
-
Have you seen a corresponding drop-off in the ListFinder pages over that time. If the canonical is kicking in, you should see some of those pages fall out as more ConsumerBase pages kick in.
Is there a reason your canonical'ing from the more indexed site to the less indexed one. It could be a mixed signal if Google things that ListFinder is a more powerful or authoritative site. Cross-domain can get tricky fast.
Unfortunately, beyond NOINDEX'ing, it's about your best option, and certainly one of your safest. It's really hard to predict what the combo of cross-domain canonical plus link would do. From a dupe content standpoint, it's risk free. From the standpoint of creating 80K links from one of your sites to another of your sites, it's a little risky (don't want to look like a link network). Since you're only talking two sites, though, it's probably not a huge issue, especially with the canonical already in place.
Google interprets cross-domain canonical heavily, so it can be a little hard to predict and control. Interestingly, the ConsumerBase site has higher Domain Authority, but the page you provided has lower Page Authority than its "sister" page. Might be a result of your internal linking structure giving more power to the ListFinder pages.
-
Great post Peter.
Here are some links of a product that is on both sites. Hopefully this will help you provide some more insight.
http://www.consumerbase.com/mailing-lists/shutterbugsphotography-enthusiasts-mailing-list.html
http://www.listfinder.com/mailing-lists/shutterbugsphotography-enthusiasts-mailing-list.htmlThe ListFinder pages are currently mostly indexed (70k out of 80k) which makes me think they are different enough from one another to not warrant a penalty.
The ConsumerBase pages started indexing well when we added the rel canonical code to LF (went from about 2k pages to 30k in early December, but since 1/2/2013 we have seen a dropoff in indexed pages down to about 5k.
Thanks!
-
With products, it's a bit hard to say. Cross-domain canonical could work, but Google can be a bit finicky about it. Are you seeing the pages on both sides in the Google index, or just one or the other? Sorry, it's a bit hard to diagnose without seeing a sample URL.
If this were more traditional syndicated content, you could set a cross-domain canonical and link the copy back to the source. That would provide an additional signal of which site should get credit. With your case, though, I haven't seen a good example of that - I don't think it would be harmful, though (to add the link, that is).
If you're talking about 80K links, then you've got 80K+ near-duplicate product pages. Unfortunately, it could go beyond just having one or the other version get filtered out. This could trigger a Panda or Panda-like penalty against the site in general. The cross-domain canonical should help prevent this, whereas the links probably won't. I do think it's smart to be proactive, though.
Worst case, you could META NOINDEX the product pages on one site - they'd still be available to users, but wouldn't rank. I think the cross-domain canonical is probably preferable here, but if you ran into trouble, META NOINDEX would be the more severe approach (and could help solve that trouble).
-
Yes, sir - that would be correct.
www.consumerbase.com and www.listfinder.com.
The sites are not 100% identical, just the content on the product pages.
-
are these two sites on the same root domain? it seems like most of the feedback you're getting are from people who are assuming they are however, it sounds to me like there are two separate domains
-
Zora,
Google accepts cross domain canonical as long as the pages have more similar content.
It is not necessary to add hyperlink pointing to canonical page. If your sites are crawler friendly, canonical hints will change search results very quickly.
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=35769
Ensure that Google doesn't find any issue with your Sitemaps. If you add products frequently, submit the updated Sitemap following the same schedule.
All the best.
-
I am sorry i am not understanding why you need a rel = in this matter if the sites are two different sites?
What is your end goal ?
-
We chose rel canonical because we still want users to be able to visit and navigate through site 2.
They are both e-commerce sites with similar products, not exactly identical sites.
-
Zora. Totally understand, but my input and what Majority of people do is redirect the traffic.
A server side htaccess 301 Redirect is your BEST choice here.
Why dont you want o use a 301 and prefer a Rel, curious on what your take is on this.
and Thanks for the rel update info i didnt know
-
Thanks for the info Hampig, I'll definitely take a look.
Rel Canonical actually works cross domain now, Google updated it from when it originally came out.
-
Zora hope you are doing well.
I came across this video about a few weeks ago. I think this is suppose to be found under Webmaster tools although i have not used it, i think it might be the best solution to get googles attention to portions of the pages and what they are suppose to be
http://www.youtube.com/watch?v=WrEJds3QeTw
Ok but i am confused a bit. You have two different domains ?
or two version of the same domain?
Because from the sound of it you have two different domains and using rel = con wont work and you would have to do a 301 redirect. Even for my sites when i change the pages around i use 301 redirect for the same existing site.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Shall we add engaging and useful FAQ content in all our pages or rather not because of duplication and reduction of unique content?
We are considering to add at the end of alll our 1500 product pages answers to the 9 most frequently asked questions. These questions and answers will be 90% identical for all our products and personalizing them more is not an option and not so necessary since most questions are related to the process of reserving the product. We are convinced this will increase engagement of users with the page, time on page and it will be genuinely useful for the visitor as most visitors will not visit the seperate FAQ page. Also it will add more related keywords/topics to the page.
Intermediate & Advanced SEO | | lcourse
On the downside it will reduce the percentage of unique content per page and adds duplication. Any thoughts about wether in terms of google rankings we should go ahead and benefits in form of engagement may outweight downside of duplication of content?0 -
Duplicate Content with URL Parameters
Moz is picking up a large quantity of duplicate content, consists mainly of URL parameters like ,pricehigh & ,pricelow etc (for page sorting). Google has indexed a large number of the pages (not sure how many), not sure how many of them are ranking for search terms we need. I have added the parameters into Google Webmaster tools And set to 'let google decide', However Google still sees it as duplicate content. Is it a problem that we need to address? Or could it do more harm than good in trying to fix it? Has anyone had any experience? Thanks
Intermediate & Advanced SEO | | seoman100 -
Canonical page 1 and rel=next/prev
Hi! I'm checking a site that has something like a News section, where they publish some posts, quite similar to a blog.
Intermediate & Advanced SEO | | teconsite
They have a canonical url pointing to the page=1. I was thinking of implementing the rel=next/ prev and the view all page and set the view all page as the canonical. But, as this is not a category page of an ecommerce site, and it would has more than 100 posts inside in less than a year, It made me think that maybe the best solution would be the following Implementing rel=next/prev
Keep page 1 as the canonical version. I don't want to make the users wait for a such a big page to load (a view all with more than 100 elements would be too much, I think) What do you think about this solution? Thank you!0 -
SEO effect of content duplication across hub of sites
Hello, I have a question about a website I have been asked to work on. It is for a real estate company which is part of a larger company. Along with several other (rival) companies it has a website of property listings which receives a feed of properties from a central hub site - so lots of potential for page, title and meta content duplication (if if isn't already occuring) across the whole network of sites. In early investigation I don't see any of these sites ranking very well at all in Google for expected search phrases. Before I start working on things that might improve their rankings, I wanted to ask some questions from you guys: 1. How would such duplication (if it is occuring) effect the SEO rankings of such sites individually, or the whole network/hub collectively? 2. Is it possible to tell if such a site has been "burnt" for SEO purposes, especially if or from any duplication? 3. If such a site or the network has been totally burnt, are there any approaches or remedies that can be made to improve the site's SEO rankings significantly, or is the only/best option to start again from scratch with a brand new site, ensuring the use of new meta descriptions and unique content? Thanks in advance, Graham
Intermediate & Advanced SEO | | gmwhite9991 -
Category Pages For Distributing Authority But Not Creating Duplicate Content
I read this interesting moz guide: http://moz.com/learn/seo/robotstxt, which I think answered my question but I just want to make sure. I take it to mean that if I have category pages with nothing but duplicate content (lists of other pages (h1 title/on-page description and links to same) and that I still want the category pages to distribute their link authority to the individual pages, then I should leave the category pages in the site map and meta noindex them, rather than robots.txt them. Is that correct? Again, don't want the category pages to index or have a duplicate content issue, but do want the category pages to be crawled enough to distribute their link authority to individual pages. Given the scope of the site (thousands of pages and hundreds of categories), I just want to make sure I have that right. Up until my recent efforts on this, some of the category pages have been robot.txt'd out and still in the site map, while others (with different url structure) have been in the sitemap, but not robots.txt'd out. Thanks! Best.. Mike
Intermediate & Advanced SEO | | 945010 -
Penalized for Duplicate Page Content?
I have some high priority notices regarding duplicate page content on my website www.3000doorhangers.com Most of the pages listed here are on our sample pages: http://www.3000doorhangers.com/home/door-hanger-pricing/door-hanger-design-samples/ On the left side of our page you can go through the different categories. Most of the category pages have similar text. We mainly just changed the industry on each page. Is this something that google would penalize us for? Should I go through all the pages and use completely unique text for each page? Any suggestions would be helpful Thanks! Andrea
Intermediate & Advanced SEO | | JimDirectMailCoach0 -
How use Rel="canonical" for our Website
How is the best way to use Rel="canonical" for our website www.ofertasdeemail.com.br, for we can say goodbye for duplicated pages? I appreciate for every help. I also hope to contribute to the SEOmoz community. Sincerely,
Intermediate & Advanced SEO | | ZZNINTERNETMEDIAGROUP
Amador Goncalves0 -
How to Remove Joomla Canonical and Duplicate Page Content
I've attempted to follow advice from the Q&A section. Currently on the site www.cherrycreekspine.com, I've edited the .htaccess file to help with 301s - all pages redirect to www.cherrycreekspine.com. Secondly, I'd added the canonical statement in the header of the web pages. I have cut the Duplicate Page Content in half ... now I have a remaining 40 pages to fix up. This is my practice site to try and understand what SEOmoz can do for me. I've looked at some of your videos on Youtube ... I feel like I'm scrambling around to the Q&A and the internet to understand this product. I'm reading the beginners guide.... any other resources would be helpful.
Intermediate & Advanced SEO | | deskstudio0