Duplicate Content within Website - problem?
-
Hello everyone,
I am currently working on a big site which sells thousands of widgets. However each widget has ten sub widgets (1,2,3... say)
My strategy with this site is to target the long tail search so I'm creating static pages for each possibly variation.
So I'll have a main product page on widgets in general, and also a page on widget1, page on widget2 etc etc.
I'm anticipating that because there's so much competition for searches relating to widgets in general, I'll get most of my traffic from people being more specific and searching for widget1 or widget 7 etc.
Now here's the problem - I am getting a lot of content written for this website - a few hundred words for each widget. However I can't go to the extreme of writing unique content for each sub widget - that would mean 10's of 1,000's of articles.
So... what do I do with the content. Put it on the main widget page was the plan but what do I do about the sub pages. I could put it there and it would make perfect sense to a reader and be relevant to people specifically looking for widget1, say, but could there be a issue with it being viewed as duplicate content.
One idea was to just put a snippet (first 100 words) on each sub page with a link back to the main widget page where the full copy would be.
Not sure whether I've made myself clear at all but hopefully I have - or I can clarify.
Thanks so much in advance
David
-
What's wrong with having ten brass widgets in ten different colors and ten buy buttons all listed on a single page?
I do that I we see lots of people buying a brass widget in every color. I think that this is great for getting more sales. If I was a shopper it would be a real frustration to visit ten pages to get one of each color - or just visit all of those pages to see which color I like best.
Most important, Google might see that and say.... This page has brass widgets in EVERY FREEKING COLOR! and decide to show it to visitors who search for them.
Now, if you are compulsive about having one page per widget and having your writer create yada yada yada content for all of them, keep in mind that you are wasting a lot of money on near duplicate content, boring your writers and spreading your pagerank out over a lot of pages.
-
David, the sub-pages as far as Goggle was concerned fed all the juice to the product page.
No the subpages were not indexed as we told Google they all came from the same page in the canonical.
How do you describe a red widget1 differently to blue widget1? The item is the same but there is only one word different in the content, so we decided to skip a physically different url for the different colours and just use different anchors on the thumbnail images. The title and alt tags would contain specific information about the colour of the widget.
If someone searches for red widget1 and we have keyword strength in widget1 they will get to the widget1 page where they will see the red widget1 and any other colours for that widget1.
The canonical allows you to specify the content origin. So if you have /category/widget1/red and /category/widget1/blue describing the same content you could use /category/widget1 in the canonical ref and both pages would give juice to the main page and get no duplicate content penality.
This only works if you have a small number of variants on each widget as Ryan pointed out, such as size, colour variations etc. Otherwise it is too confusing for humans to follow.
With the amount of content you are looking at, it is probably worthwhile getting a usability study done.
-
SEO = Manipulation doesn't it?
You can call me naive but those days of SEO are either gone or disappearing fast.
I view SEO as working to understand the ever-changing metrics search engines use to rank search results, then applying that knowledge to websites.
We are manipulated into improving our sites to provide a better user experience. The changes we make have lasting value. Other forms of SEO are always one update away from making a post asking "what happened to my site's rankings?"
-
Thanks for the replies, guys.
Oznappies - did that structure mean that all your subproduct pages were pretty much devoid of link juice? Were they even indexed? The big question is if someone searhed for 'red product a' which page showed up? Excuse my ignorance re the canonical stuff.
Ryan, Yes you are right to some degree. I am reverse engineering the website so to speak. But nevertheless I plan to offer huge value to visitors - I have spared little expense with the content writing, usability etc plus we have some fairly radical ideas that should be hugely popular with the visitors.
But I take exception that this is the wrong way to go about it. SEO = Manipulation doesn't it? The old adage 'Just make great content then users will find it and link to it and you'll dominate the serps' is a great theory but we all know in practise it doesn't work like that in 99% of the cases. To get your great product out there you have to give it a push, find an angle to exploit and this targeting of long tail is my angle.
It will be a great site I assure you
-
If the widgets are truly different products, then they should have separate product pages. If you have a weather widget, a currency exchange widget, a local time widget, etc. then you can clearly build unique content for each page.
If you offer a widget in different colors, sizes, etc. but it is really the same widget, you can't effectively generate new content for each page. Your best approach is creating a single, strong page for the widget. The "blue", "yellow" and other widget pages should be canonicalized to the main widget page.
I am getting a lot of content written for this website - a few hundred words for each widget. However I can't go to the extreme of writing unique content for each sub widget - that would mean 10's of 1,000's of articles.
That sums it up pretty well. You are having content "written" which often means it is not quality content. You are not willing to write unique content for each sub widget either. You are not developing your site for the best user experience, but instead to manipulate search engine traffic. Google is focused on preventing you from doing exactly what you are trying to do. Even if you succeed, you will be back here in a couple months asking "why did my site drop so far" after Google makes an update to adjust for this type of manipulation.
You have two options. Condense all your content to one widget page, or develop each widget page as if it was the only page on your website. When you sit down and think "I have 10k pages and I need to have content on all of them" your content will be inferior to other sites, and your SERP will reflect as much.
-
We had a similar issue but not to that scale. We had product A in Red, Blue, Green etc the first approach we used a url /category/product?id=subproduct and set id as a parameter in Google Webmaster Tools site config. This passed all the link juice to /category/product and ensured that all pages had the appropriate for the link juice page.
We then decided that all those page loads just to basically show an image for each subproduct were a pain for the customer and so decided to show small images on the /category/product page an use a jquery call to overlay a larger image when the customer clicked a particular product. This produced faster load time and better customer experience.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Internal link is creating duplicate content issues and generating 404s from website crawl.
Not sure what the best way to describe it but the site is built with Elementor page builder. We are finding out that a feature that is included with a pop modal window renders an HTML code as so: Click So when crawled I think the crawling is linking itself for some reason so the crawl returns something like this: xyz.com/builder/listing/ - what we want what we don't want xyz.com/builder/listing/ xyz.com/builder/listing/%23elementor-action%3Aaction%3Dpopup%3Aopen%26settings%3DeyJpZCI6Ijc2MCIsInRvZ2dsZSI6ZmFsc2V9/ xyz.com/builder/listing/%23elementor-action%3Aaction%3Dpopup%3Aopen%26settings%3DeyJpZCI6Ijc2MCIsInRvZ2dsZSI6ZmFsc2V9//%23elementor-action%3Aaction%3Dpopup%3Aopen%26settings%3DeyJpZCI6Ijc2MCIsInRvZ2dsZSI6ZmFsc2V9/ so you'll notice how that string in the HREF is appended each time and it loops a couple times. Could I 301 this issue, what's the best way to go about handling something like this? It's causing duplicate meta descriptions/content errors for some listing pages we have. I did add a rel='nofollow' to the anchor tag with JavaScript but not sure if that'll help.
Technical SEO | | JoseG-LP0 -
404 Error Pages being picked up as duplicate content
Hi, I recently noticed an increase in duplicate content, but all of the pages are 404 error pages. For instance, Moz site crawl says this page: https://www.allconnect.com/sc-internet/internet.html has 43 duplicates and all the duplicates are also 404 pages (https://www.allconnect.com/Coxstatic.html for instance is a duplicate of this page). Looking for insight on how to fix this issue, do I add an rel=canonical tag to these 60 error pages that points to the original error page? Thanks!
Technical SEO | | kfallconnect0 -
Duplicate Content Issues
We have some "?src=" tag in some URL's which are treated as duplicate content in the crawl diagnostics errors? For example, xyz.com?src=abc and xyz.com?src=def are considered to be duplicate content url's. My objective is to make my campaign free of these crawl errors. First of all i would like to know why these url's are considered to have duplicate content. And what's the best solution to get rid of this?
Technical SEO | | RodrigoVaca0 -
Rel=canonical overkill on duplicate content?
Our site has many different health centers - many of which contain duplicate content since there is topic crossover between health centers. I am using rel canonical to deal with this. My question is this: Is there a tipping point for duplicate content where Google might begin to penalize a site even if it has the rel canonical tags in place on cloned content? As an extreme example, a site could have 10 pieces of original content, but could then clone and organize this content in 5 different directories across the site each with a new url. This would ultimately result in the site having more "cloned" content than original content. Is this at all problematic even if the rel canonical is in place on all cloned content? Thanks in advance for any replies. Eric
Technical SEO | | Eric_Lifescript0 -
Duplicate Content and URL Capitalization
I have multiple URLs that SEOMoz is reporting as duplicate content. The reason is that there are characters in the URL that may, or may not, be capitalized depending on user input. A couple examples are: www.househitz.com/Pennsylvania/Houses-for-sale www.househitz.com/Pennsylvania/houses-for-sale www.househitz.com/Pennsylvania/Houses-for-rent www.househitz.com/Pennsylvania/houses-for-rent There are currently thousands of instances of this on the site. Is this something I should spend effort to try and resolve (may not be minor effort), or should I just ignore it and move on?
Technical SEO | | Jom0 -
How to prevent duplicate content in archives?
My news site has a number of excerpts in the form of archives based on categories that is causing duplicate content problems. Here's an example with the nutrition archive. The articles here are already posts, so it creates the duplicate content. Should I nofollow/noindex this category page along with the rest and 2011,2012 archives etc (see archives here)? Thanks so much for any input!
Technical SEO | | naturalsociety0 -
Forget Duplicate Content, What to do With Very Similar Content?
All, I operate a Wordpress blog site that focuses on one specific area of the law. Our contributors are attorneys from across the country who write about our niche topic. I've done away with syndicated posts, but we still have numerous articles addressing many of the same issues/topics. In some cases 15 posts might address the same issue. The content isn't duplicate but it is very similar, outlining the same rules of law etc. I've had an SEO I trust tell me I should 301 some of the similar posts to one authoritative post on the subject. Is this a good idea? Would I be better served implementing canonical tags pointing to the "best of breed" on each subject? Or would I be better off being grateful that I receive original content on my niche topic and not doing anything? Would really appreciate some feedback. John
Technical SEO | | JSOC0 -
Aspx filters causing duplicate content issues
A client has a url which is duplicated by filters on the page, for example: - http://www.example.co.uk/Home/example.aspx is duplicated by http://www.example.co.uk/Home/example.aspx?filter=3 The client is moving to a new website later this year and is using an out-of-date Kentico CMS which would need some development doing to it in order to enable implementation of rel canonical tags in the header, I don't have access to the server and they have to pay through the nose everytime they want the slightest thing altering. I am trying to resolve this duplicate content issue though and am wondering what is the best way to resolve it in the short term. The client is happy to remove the filter links from the page but that still leaves the filter urls in Google. I am concerned that a 301 redirect will cause a loop and don't understand the behaviour of this type of code enough. I hope this makes sense, any advice appreciated.
Technical SEO | | travelinnovations0