What is considered duplicate content?
-
Hi,
We are working on a product page for bespoke camper vans: http://www.broadlane.co.uk/campervans/vw-campers/bespoke-campers . At the moment there is only one page but we are planning add similar pages for other brands of camper vans. Each page will receive its specifically targeted content however the 'Model choice' cart at the bottom (giving you the choice to select the internal structure of the van) will remain the same across all pages.
Will this be considered as duplicate content? And if this is a case, what would be the ideal solution to limit penalty risk: A rel canonical tag seems wrong for this, as there is no original item as such. Would an iFrame around the 'model choice' enable us to isolate the content from being indexed at the same time than the page?
Thanks,
Celine
-
Hi Celine,
Google is very smart at finding content these days, so I would avoid any possible ways of trying to hide it, but looking at what is there, I wouldn't worry too much.
When looking at the model choice at the bottom of the page, it is the same for a reason, and there is no way around it. However, I wouldn't think that Google would see that as duplicate content. Lists in this manner don't normally cause issues and as mentioned above, it is more often larger 'chunks' of content that causes issues.
There are other considerations that you might want to think about before releasing a lot more pages in this manner, and one if them is making sure Google won't see the pages appearing for no other reason that to draw in search traffic for particular phrases.
Keep the pages well stocked with unique relevant content and you should be good to go.
-Andy
-
A quick way to figure this out.
Copy a entire paragraph from the content in question. Paste the whole paragraph into Google search.
any close matches? Then It is duplicate.
Thanks,
- Mike Bean
-
This type of duplicate content is common on ecommerce websites, and it isn't necessarily a big problem. However, given the fact that there will be a higher percentage of duplicate content than unique content, you run the risk of some of your pages being omitted from search results for certain queries. If that is the case, searchers will see "In order to show you the most relevant results, we have omitted some entries very similar to the (# here)already displayed. If you like, you can repeat the search with the omitted results included."
This isn't really a penalty. It's just Google being efficient with their algorithm. It shouldn't be a problem for highly targeted searches, but you may lose a little search visibility for more generic searches.
My advice is to get creative and find new ways to add more unique content to your product pages. Add testimonials, user-generated reviews, camper van adventure stories, etc.
You are right that canonical tags are wrong for this situation. Using an iframe doesn't make much sense either. Google has stated that they try to associate iframe content with the page it's embedded on anyway.
Further information:
-
Hi there,
If the higher percentage of content on each page is different from any other you should be OK. However I'd be worried about producing a whole bunch of pages like this. -
According to Google: "Duplicate content generally refers to substantive blocks of content within or across domains that either completely match other content or are appreciably similar."
The example you give is a bit in between: it contains a part which is unique - however the tabulated content which would appear on each page contains more content than the unique part. Personally I don't think that these pages would be considered duplicate content. However, if you want to be on the save side you could make a separate page with all possible configurations. This would also have the advantage that you could do without the tab's (end of of 2014 John Muller indicated that hiding content under tabs is not the best seo strategy (https://www.seroundtable.com/google-hidden-tab-content-seo-19489.html).
I wouldn't go for the iframe solution - it's a bit an outdated way to present information.
Hope this helps,
Dirk
-
Hi Celine
Good news, as you haven't made all the pages yet, now is the easiest time to implement new things! : -)
The best way I would recommend is utilising HTML Semantics http://www.w3schools.com/html/html5_semantic_elements.aspYou would have your main content inside the
and any supporting but repetitive content in <inside>tags.
Hope that helps!
King Regards
Jimmy</inside>
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Same content, different languages. Duplicate content issue? | international SEO
Hi, If the "content" is the same, but is written in different languages, will Google see the articles as duplicate content?
Intermediate & Advanced SEO | | chalet
If google won't see it as duplicate content. What is the profit of implementing the alternate lang tag?Kind regards,Jeroen0 -
Are 1x Event pages considered thin content? Should they be archived or redirected?
Since past event pages will become stale after the event, should they be keep alive and archived with only a link from a couple of places (for instance the main event page and html sitemap). Or should they be "retired" and redirected to the main event page if they are really no longer needed? They would probably be considered thin content because they won't have much traffic and will have very few links pointing to them. Right? Thanks. Inquiring minds want to know... 😉
Intermediate & Advanced SEO | | cindyt-170380 -
Duplicate content - Images & Attachments
I have been looking a GWT HTML improvements on our new site and I am scratching my head on how to stop some elements of the website showing up as duplicates for Meta Descriptions and Titles. For example the blog area: <a id="zip_0-anchor" class="zippedsection_title"></a>This blog is full of information and resources for you to implement; get more traffic, more leads an /blog//blog/page/2//blog/page/3//blog/page/4//blog/page/6//blog/page/9/The page has rel canonicals on them (using Yoast Wordpress SEO) and I can't see away of stopping the duplicate content. Can anyone suggest how to combat this? or is there nothing to worry about?
Intermediate & Advanced SEO | | Cocoonfxmedia0 -
Duplicate content for hotel websites - the usual nightmare? is there any solution other than producing unique content?
Hiya Mozzers I often work for hotels. A common scenario is the hotel / resort has worked with their Property Management System to distribute their booking availability around the web... to third party booking sites - with the inventory goes duplicate page descriptions sent to these "partner" websites. I was just checking duplication on a room description - 20 loads of duplicate descriptions for that page alone - there are 200 rooms - so I'm probably looking at 4,000 loads of duplicate content that need rewriting to prevent duplicate content penalties, which will cost a huge amount of money. Is there any other solution? Perhaps ask booking sites to block relevant pages from search engines?
Intermediate & Advanced SEO | | McTaggart0 -
Trying to advise on what seems to be a duplicate content penalty
So a friend of a friend was referred to me a few weeks ago as his Google traffic fell off a cliff. I told him I'd take a look at it and see what I could find and here's the situation I encountered. I'm a bit stumped at this point, so I figured I'd toss this out to the Moz crowd and see if anyone sees something I'm missing. The site in question is www.finishlinewheels.com In Mid June looking at the site's webmaster tools impressions went from around 20,000 per day down to 1,000. Interestingly, some of their major historic keywords like "stock rims" had basically disappeared while some secondary keywords hadn't budged. The owner submitted a reconsideration request and was told he hadn't received a manual penalty. I figured it was the result of either an automated filter/penalty from bad links, the result of a horribly slow server or possibly a duplicate content issue. I ran the backlinks on OSE, Majestic and pulled the links from Webmaster Tools. While there aren't a lot of spectacular links there also doesn't seem to be anything that stands out as terribly dangerous. Lots of links from automotive forums and the like - low authority and such, but in the grand scheme of things their links seem relevant and reasonable. I checked the site's speed in analytics and WMT as well as some external tools and everything checked out as plenty fast enough. So that wasn't the issue either. I tossed the home page into copyscape and I found the site brandwheelsandtires.com - which had completely ripped the site - it was thousands of the same pages with every element copied, including the phone number and contact info. Furthering my suspicions was after looking at the Internet Archive the first appearance was mid-May, shortly before his site took the nose dive (still visible at http://web.archive.org/web/20130517041513/http://brandwheelsandtires.com) THIS, i figured was the problem. Particularly when I started doing exact match searches for text on the finishlinewheels.com home page like "welcome to finish line wheels" and it was nowhere to be found. I figured the site had to be sandboxed. I contacted the owner and asked if this was his and he said it wasn't. So I gave him the contact info and he contacted the site owner and told them it had to come down and the owner apparently complied because it was gone the next day. He also filed a DMCA complaint with Google and they responded after the site was gone and said they didn't see the site in question (seriously, the guys at Google don't know how to look at their own cache?). I then had the site owner send them a list of cached URLs for this site and since then Google has said nothing. I figure at this point it's just a matter of Google running it's course. I suggested he revise the home page content and build some new quality links but I'm still a little stumped as to how/why this happened. If it was seen as duplicate content, how did this site with no links and zero authority manage to knock out a site that ranked well for hundreds of terms that had been around for 7 years? I get that it doesn't have a ton of authority but this other site had none. I'm doing this pro bono at this point but I feel bad for this guy as he's losing a lot of money at the moment so any other eyeballs that see something that I don't would be very welcome. Thanks Mozzers!
Intermediate & Advanced SEO | | NetvantageMarketing2 -
Need help with huge spike in duplicate content and page title errors.
Hi Mozzers, I come asking for help. I've had a client who's reported a staggering increase in errors of over 18,000! The errors include duplicate content and page titles. I think I've found the culprit and it's the News & Events calender on the following page: http://www.newmanshs.wa.edu.au/news-events/events/07-2013 Essentially each day of the week is an individual link, and events stretching over a few days get reported as duplicate content. Do you have any ideas how to fix this issue? Any help is much appreciated. Cheers
Intermediate & Advanced SEO | | bamcreative0 -
How to remove duplicate content, which is still indexed, but not linked to anymore?
Dear community A bug in the tool, which we use to create search-engine-friendly URLs (sh404sef) changed our whole URL-structure overnight, and we only noticed after Google already indexed the page. Now, we have a massive duplicate content issue, causing a harsh drop in rankings. Webmaster Tools shows over 1,000 duplicate title tags, so I don't think, Google understands what is going on. <code>Right URL: abc.com/price/sharp-ah-l13-12000-btu.html Wrong URL: abc.com/item/sharp-l-series-ahl13-12000-btu.html (created by mistake)</code> After that, we ... Changed back all URLs to the "Right URLs" Set up a 301-redirect for all "Wrong URLs" a few days later Now, still a massive amount of pages is in the index twice. As we do not link internally to the "Wrong URLs" anymore, I am not sure, if Google will re-crawl them very soon. What can we do to solve this issue and tell Google, that all the "Wrong URLs" now redirect to the "Right URLs"? Best, David
Intermediate & Advanced SEO | | rmvw0 -
Duplicate Content - Panda Question
Question: Will duplicate informational content at the bottom of indexed pages violate the panda update? **Total Page Ratio: ** 1/50 of total pages will have duplicate content at the bottom off the page. For example...on 20 pages in 50 different instances there would be common information on the bottom of a page. (On a total of 1000 pages). Basically I just wanted to add informational data to help clients get a broader perspective on making a decision regarding "specific and unique" information that will be at the top of the page. Content ratio per page? : What percentage of duplicate content is allowed per page before you are dinged or penalized. Thank you, Utah Tiger
Intermediate & Advanced SEO | | Boodreaux0