Duplicate content issues, I am running into challenges and am looking for suggestions for solutions. Please help.
-
So I have a number of pages on my real estate site that display the same listings, even when parsed down by specific features and don't want these to come across as duplicate content pages. Here are a few examples:
http://luxuryhomehunt.com/homes-for-sale/lake-mary/hanover-woods.html?feature=waterfront
http://luxuryhomehunt.com/homes-for-sale/lake-mary/hanover-woods.html
This happens to be a waterfront community so all the homes are located along the waterfront. I can use a canonical tag, but I not every community is like this and I want the parsed down feature pages to get index.
Here is another example that is a little different:
http://luxuryhomehunt.com/homes-for-sale/winter-park/bear-gully-bay.html
http://luxuryhomehunt.com/homes-for-sale/winter-park/bear-gully-bay.html?feature=without-pool
http://luxuryhomehunt.com/homes-for-sale/winter-park/bear-gully-bay.html?feature=4-bedrooms
http://luxuryhomehunt.com/homes-for-sale/winter-park/bear-gully-bay.html?feature=waterfront
So all the listings in this community happen to have 4 bedrooms, no pool, and are waterfront. Meaning that they display for each of the parsed down categories. I can possible set something that if the listings = same then use canonical of main page url, but in the next case its not so simple.
So in this next neighborhood there are 48 total listings as seen at:
http://luxuryhomehunt.com/homes-for-sale/windermere/isleworth.html
and being that it is a higher end neighborhood, 47 of the 48 listings are considered "traditional listings" and while it is not exactly all of them it is 99%.
Any recommendations is appreciated greatly.
-
Endorsing Jared for the full thread/follow-up. Unfortunately, when it comes to indexing all of these pages, you can't really have your cake and eat it too in 2012. These pages do look thin to Google - honestly, when the results don't change (and I get that that's just because the filters don't always impact the search), then it starts to look like you're just spinning out duplicates to target new keywords in the header. At high volume, that could get you into trouble (and is the kind of thing Panda has targeted).
You're right, though, if you canonical these pages, they won't get indexed and ranked. These days, my gut reaction is that the trade-off is worth it. If you focus your ranking power, the core category/neighborhood/etc. pages will get more authority, you'll reduce the risks of thin content, and you'll land search users on core pages that they can use to navigate to the options they want.
There's no solution that doesn't involve a trade-off, but I think focusing your index would be a positive trade-off. Keep in mind, too, that Google isn't really that fond of search pages - ultimately, you want them indexing the core property listings. The key is to have clear paths to those listings and to index and ranking prominent category pages. If you try to rank for every variations of ever search/sort/etc., you'll just end up diluting your ranking ability in most cases.
-
I see, and yes it will.
I know for my real estate clients, the main listings page usually ranks naturally for info that is found in listings so for example "4 bedrooms" - we have a real estate client that ranks for "x real estate" and "x homes for sale" but also ranks for "4 bedroom homes for sale in x" simply because the listings summary have number of bedrooms in them (like yours does).
However for other variables, like "no pool", its gets trickier since no one lists a house on MLS citing "no pool".
The only two ways around this are: write unique content on every main page, and include the keywords you want like 'no pool' or
write some unique content for each variable - ie write some unique copy on the "no pool" page, write some unique copy on the 'waterfront' page, etc. Even then you are still running a risk of duplicate copy. Having the titles, breadcrumbs and h1's dynamically change just might not be enough. I would put all of my efforts (including linkbuilding) to the main landing page and just make sure to include the keywords i want (thats just an opinion).
What is the data showing now - are you being penalized? Are you ranking for any "without pool" or "waterfront" terms and if so, are they getting traffic?
-
First, thanks again for responding. The challenge I have with using the canonical tag for the variable pages is that, won't it prevent google from indexing the variable pages that include some terms/ phrases I am trying to rank for?
Like Hanover Woods foreclosure homes for sale or Hanover 4 bedroom homes for sale
-
Hi Joshua,
There are a number of ways to stop Google from counting your dynamic urls as duplicates. Its unclear from your question why you can't use canonical tags for this. If you went here:
http://luxuryhomehunt.com/homes-for-sale/lake-mary/hanover-woods.html
And add the canonical tag in the HEAD section:
It will solve your issue of duplication when people choose property variables like waterfront or bedroom #. I think you were trying to point out the reason this wont work at the end of your question but Im not exactly sure what you are eluding to there?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Home Page Blog Snippets - Duplicate Content Help?
Afternoon Folks- I have been asked to contribute to a new site that has a blogfeed prominently displayed on the home page. It's laid out like this: Logo | Menu HOME PAGE SLIDER Blog 1 Title about 100 words of blog 1 Text Blog 2 Title about 100 words of blog 2 Text Blog 3 Title about 100 words of blog 3 Text Footer: -- This seems like an obvious duplicate content situation but also a way I have seen a lot of blogs laid out. (I.E. With blog content snippets being a significant portion of the home page content) I want the blogs to rank and I want the home page to rank, so I don't feel like a rel canonical on the blog post's is the correct option unless I have misunderstood their purpose. Anyone have any ideas or know how this is usually handled?
Technical SEO | | CRO_first0 -
Duplicate Content Issues
We have some "?src=" tag in some URL's which are treated as duplicate content in the crawl diagnostics errors? For example, xyz.com?src=abc and xyz.com?src=def are considered to be duplicate content url's. My objective is to make my campaign free of these crawl errors. First of all i would like to know why these url's are considered to have duplicate content. And what's the best solution to get rid of this?
Technical SEO | | RodrigoVaca0 -
Duplicate Content and URL Capitalization
I have multiple URLs that SEOMoz is reporting as duplicate content. The reason is that there are characters in the URL that may, or may not, be capitalized depending on user input. A couple examples are: www.househitz.com/Pennsylvania/Houses-for-sale www.househitz.com/Pennsylvania/houses-for-sale www.househitz.com/Pennsylvania/Houses-for-rent www.househitz.com/Pennsylvania/houses-for-rent There are currently thousands of instances of this on the site. Is this something I should spend effort to try and resolve (may not be minor effort), or should I just ignore it and move on?
Technical SEO | | Jom0 -
Duplicate Content on Multinational Sites?
Hi SEOmozers Tried finding a solution to this all morning but can't, so just going to spell it out and hope someone can help me! Pretty simple, my client has one site www.domain.com. UK-hosted and targeting the UK market. They want to launch www.domain.us, US-hosted and targeting the US market. They don't want to set up a simple redirect because a) the .com is UK-hosted b) there's a number of regional spelling changes that need to be made However, most of the content on domain.com applies to the US market and they want to copy it onto the new website. Are there ways to get around any duplicate content issues that will arise here? Or is the only answer to simply create completely unique content for the new site? Any help much appreciated! Thanks
Technical SEO | | Coolpink0 -
I am Posting an article on my site and another site has asked to use the same article - Is this a duplicate content issue with google if i am the creator of the content and will it penalize our sites - or one more than the other??
I operate an ecommerce site for outdoor gear and was invited to guest post on a popular blog (not my site) for a trip i had been on. I wrote the aritcle for them and i also will post this same article on my website. Is this a dup content problem with google? and or the other site? Any Help. Also if i wanted to post this same article to 1 or 2 other blogs as long as they link back to me as the author of the article
Technical SEO | | isle_surf0 -
Multiple URLs in CMS - duplicate content issue?
So about a month ago, we finally ported our site over to a content management system called Umbraco. Overall, it's okay, and certainly better than what we had before (i.e. nothing - just static pages). However, I did discover a problem with the URL management within the system. We had a number of pages that existed as follows: sparkenergy.com/state/name However, they exist now within certain folders, like so: sparkenergy.com/about-us/service-map/name So we had an aliasing system set up whereby you could call the URL basically whatever you want, so that allowed us to retain the old URL structure. However, we have found that the alias does not override, but just adds another option to finding a page. Which means the same pages can open under at least two different URLs, such as http://www.sparkenergy.com/state/texas and http://www.sparkenergy.com/about-us/service-map/texas. I've tried pointing to the aliased URL in other parts of the site with the rel canonical tag, without success. How much of a problem is this with respect to duplicate content? Should we bite the bullet, remove the aliased URLs and do 301s to the new folder structure?
Technical SEO | | ufmedia0 -
Duplicate content across multiple domains
I have come across a situation where we have discovered duplicate content between multiple domains. We have access to each domain and have recently within the past 2 weeks added a 301 redirect to redirect each page dynamically to the proper page on the desired domain. My question relates to the removal of these pages. There are thousands of these duplicate pages. I have gone back and looked at a number of these cached pages in google and have found that the cached pages that are roughly 30 days old or older. Will these pages ever get removed from google's index? Will the 301 redirect even be read by google to be redirected to the proper domain and page? If so when will that happen? Are we better off submitting a full site removal request of the sites that carries the duplicate content at this point? These smaller sites do bring traffic on their own but I'd rather not wait 3 months for the content to be removed since my assumption is that this content is competing with the main site. I suppose another option would be to include no cache meta tag for these pages. Any thoughts or comments would be appreciated.
Technical SEO | | jmsobe0 -
E-Commerce Duplicate Content
Hello all We have an e-commerce website with approximately 3,000 products. Many of the products are displayed in multiple categories which in turn generates a different URL! 😞 Accross the entire site I have noticed that the product pages are always outranked by competitors who have lower page authority, domain authority, total links etc etc. I am convinced this is down to duplicate content issues. I understand there is no direct penalty but how would this affect our rankings? Is page rank split between all the duplicates, which in turn lowers it's ranking potential? I have looked for a way to identify duplicate content using Google analytics but i've been unsuccessful. If the duplicate content is the issue and page rank is divided am i best using canonical or 301 redirects? Sorry if this is an obvious question but If i'm correct we could see a huge improvement in rankings accross the board. Wow! Cheers Todd
Technical SEO | | toddyC0