Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
What's the best practice for handling duplicate content of product descriptions with a drop-shipper?
-
We write our own product descriptions for merchandise we sell on our website. However, we also work with drop-shippers, and some of them simply take our content and post it on their site (same photos, exact ad copy, etc...). I'm concerned that we'll loose the value of our content because Google will consider it duplicated.
We don't want the value of our content undermined... What's the best practice for avoiding any problems with Google?
Thanks,
Adam
-
I totally agree but you should be able to have another set written with great quality - The big drop shippers always rewrite manufacturer descriptions because of this issue.
- You need to decide if the gains out-ways the costs
- You need to decide if the gains out-ways the costs
-
oops, hit submit button twice..
-
Having two sets of ad copy does effectively solve the Google issue, but it creates two non-Google issues, both of which are potentially costly. For example:
-
I have to write new copy for them which costs time and money, and even then they may still not use it, which creates enforcement issues.
-
If it's substantially different copy (and possibly inferior, because let's face it, it's hard to write two sets of good compelling copy on the same item), then it may not convert as well, which means they sell less... and we sell less
I'm not saying you can't solve my original problem with this method. I'm just saying that there are some very real costs to take into consideration
-
-
Go with David's method, or a hybrid. Present them useable text and ask that they put that on their sites and if they won't then ask they they use canonical or noindex directives.
-
You could have them add a rel- canonical - But dropshippers want your content so they can rank they will not want to use it.
-
Hey guys, thanks for all the fast responses!
I thought I remembered reading something about a technical method for demonstrating to Google that your version of content is the original version. Is there a way to do that?
And yeah, we could ask them to change their behavior (or require it), but there are costs to both and I'm wondering if there's a more effective solution (such as the possibly mythical one above).
-
penalization isnt the only thing you need to worry about its a dropshipper that is stronger then you out ranking you.
-
The best way is to give your drop shippers a feed with 1 set of descriptions and your site having another set (people will still copy but much less)
-
Are these dropshippers people who have to obey by your agreements in order to continue doing business with you? Would it hurt your business to create a requirement that they either create unique content or have their pages use the noindex code to prevent google from finding the dupe?
Do most of your dropshippers get their traffic via Organic Search? Or are they using other advertising sources?
-
There is really nothing you can do because someone else is copying your description.
The only thing I can initially come up with is asking your dropshippers to not copy descriptions.
However, the content that is duplicated and might not really negatively effect your SEO. Google understands e-commerce and a lot of the times products on e-commerce sites are very similar and they do not get penalized. Another thing is that you originally created the description and Google does index according to freshness. As long as you are indexed first with the description, I don't see how Google can penalize you.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate 'meta title' issue (AMP & NON-AMP Pages)
how to fix duplicate meta title issue in amp and non-amp pages? example.com
On-Page Optimization | | 21centuryweb
example.com/amp We have set the 'meta title' in desktop version & we don't want to change the title for AMP page as we have more than 10K pages on the website. ----As per SEMRUSH Tool---- ABOUT THIS ISSUE It is a bad idea to duplicate your title tag content in your first-level header. If your page’s <title>and <h1> tags match, the latter may appear over-optimized to search engines. Also, using the same content in titles and headers means a lost opportunity to incorporate other relevant keywords for your page.</p> <p><strong>HOW TO FIX IT</strong></p> <p>Try to create different content for your <title> and <h1> tags.<br /><br />this is what they are recommending, for the above issue we have asked our team to create unique meta and post title for desktop version but what about AMP page?<br /><br />Please help!</p></title>0 -
Image naming best practices?
While I have found many good sources of information for naming images for SEO purposes, I'm having trouble finding an up-to-date, exhaustive and authoritative source for image names, alt tags, etc. For instance... Max characters for image name? Max hyphens? How descriptive should you be? "ice-cream-flavors-icon_._jpg" or "ice-cream-flavors.jpg" or simply "ice-cream.jpg" How similar should the image name, alt text and page title be? At what point are you overusing a keyword? Rules to follow? So much more, but you get the idea! Anyone have a good reference or an answer to all things related to images and SEO? Thanks!
On-Page Optimization | | OSD0 -
Duplicate Content on Event Pages
My client has a pretty popular service of event listings and, in hope of gathering more events, they opened up the platform to allow users to add events. This works really well for them and they are able to garner a lot more events this way. The major problem I'm finding is that many event coordinators and site owners will take the copy from their website and copy and paste it, duplicating a lot of the content. We have editor picks that contain a lot of unique content but the duplicate content scares me. It hasn't hurt our page ranking (we have a page ranking of 7) but I'm wondering if this is something that we should address. We don't have the manpower to eliminate all the duplication but if we cut down the duplication would we experience a significant advantage over people posting the same event?
On-Page Optimization | | mattdinbrooklyn0 -
Solve duplicate content issues by using robots.txt
Hi, I have a primary website and beside that I also have some secondary websites with have same contents with primary website. This lead to duplicate content errors. Because of having many URL duplicate contents, so I want to use the robots.txt file to prevent google index the secondary websites to fix the duplicate content issue. Is it ok? Thank for any help!
On-Page Optimization | | JohnHuynh0 -
Does schema.org assist with duplicate content concerns
The issue of duplicate content has been well documented and there are lots of articles suggesting to noindex archive pages in WordPress powered sites. Schema.org allows us to mark-up our content, including marking a components URL. So my question simply, is no-indexing archive (category/tag) pages still relevant when considering duplicate content? These pages are in essence a list of articles, which can be marked as an article or blog posting, with the url of the main article and all the other cool stuff the scheme gives us. Surely Google et al are smart enough to recognise these article listings as gateways to the main content, therefore removing duplicate content concerns. Of course, whether or not doing this is a good idea will be subjective and based on individual circumstances - I'm just interested in whether or not the search engines can handle this appropriately.
On-Page Optimization | | MarkCA0 -
Percentage of duplicate content allowable
Can you have ANY duplicate content on a page or will the page get penalized by Google? For example if you used a paragraph of Wikipedia content for a definition/description of a medical term, but wrapped it in unique content is that OK or will that land you in the Google / Panda doghouse? If some level of duplicate content is allowable, is there a general rule of thumb ratio unique-to-duplicate content? thanks!
On-Page Optimization | | sportstvjobs0 -
Avoiding "Duplicate Page Title" and "Duplicate Page Content" - Best Practices?
We have a website with a searchable database of recipes. You can search the database using an online form with dropdown options for: Course (starter, main, salad, etc)
On-Page Optimization | | smaavie
Cooking Method (fry, bake, boil, steam, etc)
Preparation Time (Under 30 min, 30min to 1 hour, Over 1 hour) Here are some examples of how URLs may look when searching for a recipe: find-a-recipe.php?course=starter
find-a-recipe.php?course=main&preperation-time=30min+to+1+hour
find-a-recipe.php?cooking-method=fry&preperation-time=over+1+hour There is also pagination of search results, so the URL could also have the variable "start", e.g. find-a-recipe.php?course=salad&start=30 There can be any combination of these variables, meaning there are hundreds of possible search results URL variations. This all works well on the site, however it gives multiple "Duplicate Page Title" and "Duplicate Page Content" errors when crawled by SEOmoz. I've seached online and found several possible solutions for this, such as: Setting canonical tag Adding these URL variables to Google Webmasters to tell Google to ignore them Change the Title tag in the head dynamically based on what URL variables are present However I am not sure which of these would be best. As far as I can tell the canonical tag should be used when you have the same page available at two seperate URLs, but this isn't the case here as the search results are always different. Adding these URL variables to Google webmasters won't fix the problem in other search engines, and will presumably continue to get these errors in our SEOmoz crawl reports. Changing the title tag each time can lead to very long title tags, and it doesn't address the problem of duplicate page content. I had hoped there would be a standard solution for problems like this, as I imagine others will have come across this before, but I cannot find the ideal solution. Any help would be much appreciated. Kind Regards5