What is considered duplicate content?
-
Hi,
We are working on a product page for bespoke camper vans: http://www.broadlane.co.uk/campervans/vw-campers/bespoke-campers . At the moment there is only one page but we are planning add similar pages for other brands of camper vans. Each page will receive its specifically targeted content however the 'Model choice' cart at the bottom (giving you the choice to select the internal structure of the van) will remain the same across all pages.
Will this be considered as duplicate content? And if this is a case, what would be the ideal solution to limit penalty risk: A rel canonical tag seems wrong for this, as there is no original item as such. Would an iFrame around the 'model choice' enable us to isolate the content from being indexed at the same time than the page?
Thanks,
Celine
-
Hi Celine,
Google is very smart at finding content these days, so I would avoid any possible ways of trying to hide it, but looking at what is there, I wouldn't worry too much.
When looking at the model choice at the bottom of the page, it is the same for a reason, and there is no way around it. However, I wouldn't think that Google would see that as duplicate content. Lists in this manner don't normally cause issues and as mentioned above, it is more often larger 'chunks' of content that causes issues.
There are other considerations that you might want to think about before releasing a lot more pages in this manner, and one if them is making sure Google won't see the pages appearing for no other reason that to draw in search traffic for particular phrases.
Keep the pages well stocked with unique relevant content and you should be good to go.
-Andy
-
A quick way to figure this out.
Copy a entire paragraph from the content in question. Paste the whole paragraph into Google search.
any close matches? Then It is duplicate.
Thanks,
- Mike Bean
-
This type of duplicate content is common on ecommerce websites, and it isn't necessarily a big problem. However, given the fact that there will be a higher percentage of duplicate content than unique content, you run the risk of some of your pages being omitted from search results for certain queries. If that is the case, searchers will see "In order to show you the most relevant results, we have omitted some entries very similar to the (# here)already displayed. If you like, you can repeat the search with the omitted results included."
This isn't really a penalty. It's just Google being efficient with their algorithm. It shouldn't be a problem for highly targeted searches, but you may lose a little search visibility for more generic searches.
My advice is to get creative and find new ways to add more unique content to your product pages. Add testimonials, user-generated reviews, camper van adventure stories, etc.
You are right that canonical tags are wrong for this situation. Using an iframe doesn't make much sense either. Google has stated that they try to associate iframe content with the page it's embedded on anyway.
Further information:
-
Hi there,
If the higher percentage of content on each page is different from any other you should be OK. However I'd be worried about producing a whole bunch of pages like this. -
According to Google: "Duplicate content generally refers to substantive blocks of content within or across domains that either completely match other content or are appreciably similar."
The example you give is a bit in between: it contains a part which is unique - however the tabulated content which would appear on each page contains more content than the unique part. Personally I don't think that these pages would be considered duplicate content. However, if you want to be on the save side you could make a separate page with all possible configurations. This would also have the advantage that you could do without the tab's (end of of 2014 John Muller indicated that hiding content under tabs is not the best seo strategy (https://www.seroundtable.com/google-hidden-tab-content-seo-19489.html).
I wouldn't go for the iframe solution - it's a bit an outdated way to present information.
Hope this helps,
Dirk
-
Hi Celine
Good news, as you haven't made all the pages yet, now is the easiest time to implement new things! : -)
The best way I would recommend is utilising HTML Semantics http://www.w3schools.com/html/html5_semantic_elements.aspYou would have your main content inside the
and any supporting but repetitive content in <inside>tags.
Hope that helps!
King Regards
Jimmy</inside>
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate content. Competing for rank.
Scenario: An automotive dealer lists cars for sale on their website. The descriptions are very good and in depth at 1,200 words per car. However chunks of the copy are copied from car review websites and weaved into their original copy. Q1: This is flagged in copyscape - how much of an issue is this for Google? Q2: The same stock with the same copy is fed into a popular car listing website - the dealer's website and the classifieds website often rank in the top two positions (sometimes the dealer on top other times the classifieds site). Is this a good or a bad thing? Are you risking being seen as duplicating/scraping content? Thank you.
Intermediate & Advanced SEO | | Bee1590 -
Identifying Duplicate Content
Hi looking for tools (beside Copyscape or Grammarly) which can scan a list of URLs (e.g. 100 pages) and find duplicate content quite quickly. Specifically, small batches of duplicate content, see attached image as an example. Does anyone have any suggestions? Cheers. 5v591k.jpg
Intermediate & Advanced SEO | | jayoliverwright0 -
Is This Considered Duplicate Content?
My site has entered SEO hell and I am not sure how to fix it. Up until 18 months ago I had tremendous success on Google and Bing and now my website appears below my Facebook page for the term "Direct Mail Raleigh." What makes it even more frustrating is my competitors have done no SEO and they are dominating this keyword. I thought that the issue was due to harmful inbound links and two months ago I disavowed ones that were clearly spam. Somehow my site has actually gone down! I have a blog that I have updated infrequently and I do not know if it I am getting punished for duplicate content. On Google Webmaster Tools it says I have 279 crawled and indexed pages. Yesterday when I ran the MOZ crawl check I was amazed to find 1150 different webpages on my site. Despite the fact that it does not appear on the webmaster tools I have three different webpages due to the format that the Wordpress blog was created: "http://www.marketplace-solutions.com/report/part2leadershi/", "http://www.marketplace-solutions.com/report/page/91/" and "http://www.marketplace-solutions.com/report/category/competent-leadership/page/3/" What does not make sense to me is why Google only indexed 279 webpages AND why MOZ did not identify these three webpages as duplicate content with the Crawl Test Tool. Does anyone have any ideas? Would it be as easy as creating a massive robot.txt file and just putting 2 of the 3 URLs in that file? Thank you for your help.
Intermediate & Advanced SEO | | DR700950 -
PDF for link building - avoiding duplicate content
Hello, We've got an article that we're turning into a PDF. Both the article and the PDF will be on our site. This PDF is a good, thorough piece of content on how to choose a product. We're going to strip out all of the links to our in the article and create this PDF so that it will be good for people to reference and even print. Then we're going to do link building through outreach since people will find the article and PDF useful. My question is, how do I use rel="canonical" to make sure that the article and PDF aren't duplicate content? Thanks.
Intermediate & Advanced SEO | | BobGW0 -
About robots.txt for resolve Duplicate content
I have a trouble with Duplicate content and title, i try to many way to resolve them but because of the web code so i am still in problem. I decide to use robots.txt to block contents that are duplicate. The first Question: How do i use command in robots.txt to block all of URL like this: http://vietnamfoodtour.com/foodcourses/Cooking-School/
Intermediate & Advanced SEO | | magician
http://vietnamfoodtour.com/foodcourses/Cooking-Class/ ....... User-agent: * Disallow: /foodcourses ( Is that right? ) And the parameter URL: h
ttp://vietnamfoodtour.com/?mod=vietnamfood&page=2
http://vietnamfoodtour.com/?mod=vietnamfood&page=3
http://vietnamfoodtour.com/?mod=vietnamfood&page=4 User-agent: * Disallow: /?mod=vietnamfood ( Is that right? i have folder contain module, could i use: disallow:/module/*) The 2nd question is: Which is the priority " robots.txt" or " meta robot"? If i use robots.txt to block URL, but in that URL my meta robot is "index, follow"0 -
Is there a way to stop my product pages with the "show all" catagory/attribute from duplicating content?
If there were less pages with the "show all" attribute it would be a simple fix by adding the canonical URL tag. But seeing that there are about 1,000 of them I was wondering if their was a broader fix that I could apply.
Intermediate & Advanced SEO | | cscoville0 -
Cross-Domain Canonical and duplicate content
Hi Mozfans! I'm working on seo for one of my new clients and it's a job site (i call the site: Site A).
Intermediate & Advanced SEO | | MaartenvandenBos
The thing is that the client has about 3 sites with the same Jobs on it. I'm pointing a duplicate content problem, only the thing is the jobs on the other sites must stay there. So the client doesn't want to remove them. There is a other (non ranking) reason why. Can i solve the duplicate content problem with a cross-domain canonical?
The client wants to rank well with the site i'm working on (Site A). Thanks! Rand did a whiteboard friday about Cross-Domain Canonical
http://www.seomoz.org/blog/cross-domain-canonical-the-new-301-whiteboard-friday0 -
Accepting RSS feeds. Does it = duplicate content?
Hi everyone, for a few years now I've allowed school clients to pipe their news RSS feed to their public accounts on my site. The result is a daily display of the most recent news happening on their campuses that my site visitors can browse. We don't republish the entire news item; just the headline, and the first 150 characters of their article along with a Read more link for folks to click if they want the full story over on the school's site. Each item has it's own permanent URL on my site. I'm wondering if this is a wise practice. Does this fall into the territory of duplicate content even though we're essentially providing a teaser for the school? What do you think?
Intermediate & Advanced SEO | | peterdbaron0