Crawl Diagnostics Summary - Duplicate Content
-
Hello SEO Experts,
I am a developer at www.bowanddrape.com and we are working on improving the SEO of the website. The SEOMoz Crawl Diagnostics Summary shows that following 2 URL have duplicate content.
http://www.bowanddrape.com/clothing/Tan+Accessories+Calfskin+Belt/50_5142
http://www.bowanddrape.com/clothing/Black+Accessories+Calfskin+Belt/50_5143
Can you please suggest me ways to fix this problem?
Is the duplicate content error because of same "The Details", "Size Chart" and "The Silhouette" and "You may also like" ?
Thanks,
Chirag
-
It's tough, because these variations/customizations are legitimately what you do. My gut feeling, though, is that 80K (I'm seeing 90K with a site: search) indexed pages is just too much for your current link profile. It doesn't mean you'll get in trouble, but it could mean that your ranking power is spread far too thin.
While it's not a decision I'd take lightly, I do think there's an advantage here to either:
(1) Consolidating variations under one URL
(2) Having multiple URLs, but possibly using rel=canonical (I think that's your best bet) to focus Google on one parent URL for each product
-
Dr. Peter, Thanks for the useful insight, right now google web master tool shows that 82,563 pages on our website are in google's index, but sadly none are getting any direct traffic from google search results. We are "design your own dress company" so each "product" can have 1000s of variations, most are similar to google, but not to the end-user. So I think what you are saying is that consolidating all variations of 1 product to 1 page could result in more power on the single product page. Can you please confirm?
-
I'm gonna disagree mildly. It is common to have color variation pages, and it is perfectly useful to end-users. So, you're not doing anything wrong, in that sense. However, these pages don't look very different to Google (minor variations in title and content), and so we do flag them as near duplicates because Google might consider them "thin". At large scale, that could dilute your ranking ability.
If you have 100s or 1000s of these pages and a relatively weak link profile, it might be worth considering canonical tags here. The trade-off is that you would consolidate your ranking power, but one variation would fall out of search results. So, it really depends not only on the scope of the problem, but the strength of the site, and how important these long-tail color-based searches are to your current traffic. There's no one-sized-fits-all answer.
-
Thanks Eyepaq. I can keep it as is, but I will try to make them more brown or black by adding brown or black to the The Details and the The Silhouette.
-
Thanks. I will try to make them more unique.
-
At the moment the pages are too similar so are coming up as dups, (they also will most likely compete with each other in the serps too)
My advice would be either make them more different content wise, or have one page that covers both terms (I would guess they would be long tail terms anyway, so that might be the best option)
using canonical links it telling google they are the same page content wise and which is the "master page" to show in the serps
-
In this case you can let those be as they are...
No harm to the website or pages for this "issue" - it is a common think for this type of color / type differences and you should not add rel canonical or redirect it s you need them both in the search pages.
There is no down side of having those like this.
Cheers.
-
Thanks for the Reply Bryan. I have used canonical links at other places on the website, where the pages are same.
I want to make the 2 pages so that I can attract users both user searching for black belt as well as brown bag. Would adding canonical links help me in doing that, or am I thinking of this in the wrong way?
-
You need to add a canonical tags to let search engines know that the content is almost identical.
here is an awesome post to get you all set up: http://www.seomoz.org/blog/canonical-url-tag-the-most-important-advancement-in-seo-practices-since-sitemaps
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Regarding rel=canonical on duplicate pages on a shopping site... some direction, please.
Good morning, Moz community: My name is David, and I'm currently doing internet marketing for an online retailer of marine accessories. While many product pages and descriptions are unique, there are some that have the descriptions duplicated across many products. The advice commonly given is to leave one page as is / crawlable (probably best for one that is already ranking/indexed), and use rel=canonical on all duplicates. Any idea for direction on this? Do you think it is necessary? It will be a massive task. (also, one of the products that we rank highest for, we have tons of duplicate descriptions.... so... that is sort of like evidence against the idea?) Thanks!
Web Design | | DavidCiti0 -
Best Practices for Leveraging Long Tail Content & Gated Content
Our B2B site has a lot of of long form content (e.g., transcriptions from presentations and webinars). We'd like to leverage the long tail SEO traffic driven to these pages and convert those visitors to leads. Essentially, we'd like Google to index all this lengthy, keyword-rich content AND we'd like to put up a read gate that requires users to register before viewing the full article. This is a B2B site, and the goal is to generate leads. Some considerations and questions: How much of the content to share before requiring registration? Ask too soon and it's a terrible user experience, give too much away and our business objectives are not met. Design-wise, what are good ways to do this? I notice Moz uses a "teaser" to block Mozinar content, and I've seen modals and blur bars on other sites. Any gotchas that Google doesn't like that we should be aware of? Trying to avoid anything that might seem like cloaking. Is it better to split the content across several pages (split a 10K word doc across 10 URLs and include a read gate on each) or keep to one page? Thank you!
Web Design | | Allie_Williams0 -
Why would a developer build all page content in php?
Picked up a new client. Site is built on Wordpress. Previous developer built nearly all page content in their custom theme's PHP files. In other words, the theme's "page.php" file contains virtually all the HTML for each of the site's pages. Each individual page's back-end page editor appears blank, except for some of the page text. No markup, no widgets, no custom fields. And no dedicated, page-specific php files either. Pages are differentiated within page.php using: elseif (is_page("27") Has anyone ever come across this approach before? Why might someone do this?
Web Design | | mphdavidson0 -
Redirects Not Working / Issue with Duplicate Page Titles
Hi all We are being penalised on Webmaster Tools and Crawl Diagnostics for duplicate page titles and I'm not sure how to fix it.We recently switched from HTTP to HTTPS, but when we first switched over, we accidentally set a permanent redirect from HTTPS to HTTP for a week or so(!).We now have a permanent redirect going the other way, HTTP to HTTPS, and we also have canonical tags in place to redirect to HTTPS.Unfortunately, it seems that because of this short time with the permanent redirect the wrong way round, Google is confused as sees our http and https sites as duplicate content.Is there any way to get Google to recognise this new (correct) permanent redirect and completely forget the old (incorrect) one?Any ideas welcome!
Web Design | | HireSpace0 -
Subdomains, duplicate content and microsites
I work for a website that generates a high amount of unique, quality content. This website though has had development issues with our web builder and they are going to separate the site into different subdomains upon launch. It's a scholarly site so the subdomains will be like history and science and stuff. Don't ask why aren't we aren't using subdirectories because trust me I wish we could. So we have to use subdomains and I'm wondering a couple questions. Will the duplication of coding, since all subdomains will have the same design and look, heavily penalize us and is there any way around that? Also if we generate a good amount of high quality content on each site could we link all those sites to our other site as a possible benefit for link building? And finally, would footer links, linking all the subdirectories, be a good thing to put in?
Web Design | | mdorville0 -
Penalized by duplicate content?
Hello, I am in a very weird position. I am managing a website(EMD) which a part of it dynamically creates pages. The former webmaster who create this system though that this would help with SEO but I dought! The thing is that now the site has about 1500 pages which must look duplicate but are they really duplicate? Each page has a unique URL but the content is pretty much the same: one image and a different title with 5-8 words. There is more: All these pages are not accessible by the users but only for the crawlers!!! This URL machine is a part of a php - made photo gallery which i never understood the sense of it! The site overall is not performing very well in SERP, especially after Penguin, but judging by the link profile, the Domain authority, construction (ok besides that crazy photo gallery) and content, it never reached the position it should have in the past. The majority of these mysterious pages - and mostly their images - are cached by Google and some of them are in top places to some SERP - the ones that match the small title on page - but the numbers are poor, 10 - 15 clicks per month. Are these pages considered as duplicated, although they are cached, and is it safe for the site just to remove 1500 at once? The seomoz tools have pointed some of them as dups but the majority not! Can these pages impact the image of the whole site in search engines?( drop in Google and has disappeared from Yahoo and Bing!) Do I also have to tell Google about the removal? I have not seen anything like it before so any comment would be helpful! Thank you!
Web Design | | Tz_Seo0 -
How serious is duplicate page content?
We just launched our site on a new platform - Magento Enterprise. We have a wholesale catalog and and retail catalog. We have up to 3 domains pointing to each product. We are getting tons of duplicate content errors. What are the best practices for dealing with this? Here is an example: mysite.com/product.html mysite.com/category/product.html mysite.com/dynamic-url
Web Design | | devonkrusich0 -
Can google crawl text in jquery sliders?
We are redesigning our website and want to present a fair amount of text within jquery sliders. Will google crawl this text or is it treated the same way as actual script? Perhaps there is a way to just have the text as plain html but use jquery to display it?
Web Design | | Netboost0