How to Solve Duplicate Page Content Issue?
-
I have created one campaign over SEOmoz tools for my website. I have found 89 duplicate content issue from report.
Please, look in to Duplicate Page Content Issue.
I am quite confuse to resolve this issue. Can any one suggest me best solution to resolve it?
-
No probs glad to help!
Best of luck!
-
Oh great. That's fine. Now, I got idea... What's wrong with my site. If I suppose to find that duplicate page in Admin so I may not able to find that pages. Right?? So, I just need to set 301 (Permanent Redirection) with help of htaccess. Now, It's clear. Thanks a lot for your prompt reply and quick discussion on my issue.
-
Are you talking about HTML sitemap or XML sitemap?
If HTML sitemap so you are right and I suppose to make it live very soon with proper structure.
If you are talking about XML sitemap so I have created it with two partition.
http://www.vistastores.com/main_sitemap.xml
http://www.vistastores.com/products_sitemap.xml
So, What you think about it?
301 redirection is final. Right?? After removal of duplicate pages.
-
The tool you used to find these pages seems to work fine ;).
It's simply a crawl so use the seomoz data as it's a crawl of your site
Set a 301 redirect with the complete list of duplicated content, you don't need to delete the pages(it's not deleted anyway since it's dynamically created pages)
/ G
-
Ok! Â Then we have that cleared
If you already are using URL rewrite there shouldn't be any duplicated content in the manner the list shows.
- But since there is a problem I would check so that you're not using a sitemap that dynamically crawls and creates these urls for the pages.
If that's not the issue:
- We then come back to doing the 301:s by using .htaccess.
/ Gustav
-
In addition you may want to remove "zero products" pages from index and not link to them as they are not good for users or search engines.
Yes, you are right. I want to remove zero product pages as well as all pages which were created due to human error.
As I mentioned above: Following page is not available on website for buyers. Buyers will not able to see this page any more because, not a single page available on website which is available with hyperlink to that page.
http://www.vistastores.com/126/cookwares.html
But, SEOmoz crawler detect it and when I see in excel so I quite confuse. Because, that page is working and visible with all products which are available to original page.
Users are not able to go that duplicate page but Google crawler can go there and able to detect as duplication.
I have bit knowledge about duplication and assume that Google will detect both page as duplicate. That's it. And, I want to resolve it.
-
I think, This issue is not regarding URL rewriting. I can say very sure for it. As I said, duplicate page content URLs are not available on website. If any visitors will visit website and go to each page so that visitor will not able to find duplicate page.
URL rewriting will fix issue regarding URL structure but, what about duplication which is available in my own website to compete both pages with same keyword?
-
hi again!
After reading you follow up question I know a better solution for you
Instead of going canonical which takes much manual labor or doing 301:s since both are simply just temporary fixes for the problem and not the cure.
The best way and what I always recommend our clients is working with URL rewrite. This takes care of this problem completely, however it takes some coding to implement.
If you can handle this yourself and only need hints and guides read below:
- I assume that you use Apache server?
- If so read this:Â http://httpd.apache.org/docs/2.0/misc/rewriteguide.html
Or:
- Contact the admin for the webshop and ask them to make an URL Rewrite based on the URL structure that you would want to use
- Good structure: Â root/productcategory/Product
Hope this helped you
-
I am quite confuse with 301 redirect. There are too many duplicate pages which were generated due to wrong categorization or create new same category rather edit old category.
I want to remove all duplicate pages from my website and want to set 301 redirect.
I want to give one example for it.
Original Page:
http://www.vistastores.com/125_126/kitchen-and-dining/cookwares.html
Duplicate Page:
http://www.vistastores.com/126/cookwares.html
Duplicate pages were created due to Admin issue or by human error. No one will find our land on duplicate page during website surfing.
Now, I want to detect and remove all pages which are available with duplicate content.
So, Does it matter to delete that pages from website and set 301 to associated page or home page?
-
Best practice in your case would be to implement URL canonicalisation (rel="canonical").
Watch this:Â http://www.google.com/support/webmasters/bin/answer.py?answer=139394Â (contains explanation and examples)
In addition to this try to prevent page duplication form happening in the first place, though this may need to be done on a programming level.
It seems that string such as "6_129_130" appears based on the category and navigational path. For example if user browses from outdoor and home decor and arrive on the same page the URL will have different number.
In addition you may want to remove "zero products" pages from index and not link to them as they are not good for users or search engines.
-
Hi!
Good question, one I often come across
I would say this is not a real issue for you in the serp. But if you want to fix it and competition is fierce in your field you should of course do it
Step one: Look up your index at google: simply with site:"youdomain.com" and see if there are any duplicated content in the index(chance is slim)
Step two: look up your webshop cms, perhaps there is a function for redirecting dynamic pages to a static page(usually there is)
Step three: If it's not possible to do the 301:s from webshop admin use the .htaccess file and implement 301:s with the url:s you have in the list.
Other comments:
If you would like to clean up these URLs work with 301:s and decide which structure is most important.(base it on the current index)
There are several ways to do an 301: but in this case I would say the easiest way since you have the list is to go to the .htacces and put the 301 redirect list there.
This way could of course be bad for you if the duplicated pages are somehow necessary for the webshop so start to look at the webshop to make certain that you don't mess up any critical part by doing it by .htaccess.
Best regards!
/ Gustav
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Email and landing page duplicate content issue?
Hi Mozers, my question is, if there is a web based email that goes to subscribers, then if they click on a link it lands on a Wordpress page with very similar content, will Google penalize us for duplicate content? If so is the best workaround to make the email no index no follow? Thanks!
Technical SEO | | CalamityJane770 -
Looking at creating some auto-generated pages - duplicate content?
Hi Everyone! We just launched a new version of our research site and the main CTA on the page sends users to a subdomain that's blocked by robots.txt. The subdomain link is one of our PPC landing pages and they would be duplicate content for every model (cars). We're also looking at a new content stream of deals pages, on the main domain. The thought process was that we could rank these pages for things like "Volkswagen golf deals" and also use them as canonical URLs from the PPC pages so that Panda doesn't get mad at us for sending hundreds of links to a subdomain that's blocked. It's going to take us a lot of time to write the copy for the deals pages, so if we auto-generate it by pulling a paragraph of copy from the car review, and numerical stats about that model, will it be classes as duplicate and/or is there any downside to doing it? Review Page:Â http://www.carwow.co.uk/car-reviews/Ford/Fiesta Deals Page:Â http://www.carwow.co.uk/deals/Ford/Fiesta PPC Landing Page:Â http://quotes.carwow.co.uk/buy/Ford/Fiesta I can't help but feel that this may all be a bit overkill and perhaps it makes more sense to build 1 central deals page per model with unique content that we can also send the PPC traffic to, then life any block from the quotes. subdomain. But that will take time and we'd also like a quick solution. I'd also question if it's even an issue to link to a blocked subdomain, Google adds the quote URL into the index but can't crawl it, which I've been told is bad - but is it bad enough to do something about? Thanks, JP
Technical SEO | | Matt.Carwow0 -
Https Duplicate Content
My previous host was using shared SSL, and my site was also working with https which I didnât notice previously. Now I am moved to a new server, where I donât have any SSL and my websites are not working with https version. Problem is that I have found Google have indexed one of my blog http://www.codefear.com with https version too. My blog traffic is continuously dropping I think due to these duplicate content. Now there are two results one with http version and another with https version. I searched over the internet and found 3 possible solutions. 1 No-Index https version
Technical SEO | | RaviAhuja
2 Use rel=canonical
3 Redirect https versions with 301 redirection Now I donât know which solution is best for me as now https version is not working. One more thing I donât know how to implement any of the solution. My blog is running on WordPress. Please help me to overcome from this problem, and after solving this duplicate issue, do I need Reconsideration request to Google. Thank you0 -
SEOMOZ and non-duplicate duplicate content
Hi all, Looking through the lovely SEOMOZ report, by far its biggest complaint is that of perceived duplicate content. Its hard to avoid given the nature of eCommerce sites that oestensibly list products in a consistent framework. Most advice about duplicate content is about canonicalisation, but thats not really relevant when you have two different products being perceived as the same. Thing is, I might have ignored it but google ignores about 40% of our site map for I suspect the same reason. Basically I dont want us to appear "Spammy". Actually we do go to a lot of time to photograph and put a little flavour text for each product (in progress). I guess my question is, that given over 700 products, why 300ish of them would be considered duplicates and the remaning not? Here is a URL and one of its "duplicates" according to the SEOMOZ report: http://www.1010direct.com/DGV-DD1165-970-53/details.aspx
Technical SEO | | fretts
http://www.1010direct.com/TDV-019-GOLD-50/details.aspx Thanks for any help people0 -
We have set up 301 redirects for pages from an old domain, but they aren't working and we are having duplicate content problems - Can you help?
We have several old domains. One is http://www.ccisound.com - Our "real" site is http://www.ccisolutions.com The 301 redirect from the old domain to the new domain works. However, the 301-redirects for interior pages, like: http://www.ccisolund.com/StoreFront/category/cd-duplicators do not work. This URL should redirect to http://www.ccisolutions.com/StoreFront/category/cd-duplicators but as you can see it does not. Our IT director supplied me with this code from the HT Access file in hopes that someone can help point us in the right direction and suggest how we might fix the problem: RewriteCond%{HTTP_HOST} ccisound.com$ [NC] RewriteRule^(.*)$ http://www.ccisolutions.com/$1 [R=301,L] Any ideas on why the 301 redirect isn't happening? Thanks all!
Technical SEO | | danatanseo0 -
Duplicate content
I'm getting an error showing that two separate pages have duplicate content. Â The pages are: | Help System: Domain Registration Agreement - Registrar Register4Less, Inc. http://register4less.com/faq/cache/11.html 1 27 1 Help System: Domain Registration Agreement - Register4Less Reseller (Tucows) http://register4less.com/faq/cache/7.html | These are both registration agreements, one for us (Register4Less, Inc.) as the registrar, and one for Tucows as the registrar. Â The pages are largely the same, but are in fact different. Â Is there a way to flag these pages as not being duplicate content? Thanks, Doug.
Technical SEO | | R4L0 -
Duplicate content + wordpress tags
According to SEOMoz platform, one of my wordpress websites deals with duplicate content because of the tags I use. How should I fix it? Is it loyal to remove tag links from the post pages?
Technical SEO | | giankar0 -
Is 100% duplicate content always duplicate?
Bit of a strange question here that would be keen on getting the opinions of others on. Let's say we have a web page which is 1000 lines line, pulling content from 5 websites (the content itself is duplicate, say rss headlines, for example). Obviously any content on it's own will be viewed by Google as being duplicate and so will suffer for it. However, given one of the ways duplicate content is considered is a page being x% the same as another page, be it your own site or someone elses. In the case of our duplicate page, while 100% of the content is duplicate, the page is no more than 20% identical to another page so would it technically be picked up as duplicate. Hope that makes sense? My reason for asking is I want to pull latest tweets, news and rss from leading sites onto a site I am developing. Obviously the site will have it's own content too but also want to pull in external.
Technical SEO | | Grumpy_Carl0