Why Does SEOMOZ Crawl show that i have 5,769 pages with Duplicate Content
-
Hello...
I'm trying to do some analysis on my site (http://goo.gl/JgK1e) and SEOMOZ Crawl Diagnostics is telling me that I have 5,769 pages with duplicate content.
Can someone, anyone, please help me understand:
-
how does SEOMOZ determine if i have duplicate content
-
Is it correct ? Are there really that many pages of duplicate content
-
How do i fix this, if true <---- ** Most important **
Thanks in advance for any help!!
-
-
Looks like its now sorted!
Checked link: http://www.plasticstorage.com/
Type of redirect: 301 Moved Permanently
Redirected to: http://plasticstorage.com/
-
Hey Lavellester,
I believe we got our 302/301 issue resolved. Any way you can check through whatever tool you were using last week?
We are still trying to figure out how to fix the code creating the dupe content, but can you advise if i fixed the redirect properly?
Thanks,
-
it will be related to the code/logic of your site creating the duplicate pages. You could work out why/update the code so you only have 1 page for each product OR you could use the rel canonical tag to resolve the issue.
Just thought...as you appear to have so many duplicate pages it may be quicker to look at the logic of the site and fix it all in one go.
-
Lavellester, thanks for all of your help, i am going to have that redirect changed to a 301 ASAP.
Do you have any idea where that duplicate content is coming from since i only have that item in my database once?
Thanks
-
The redirect is still a 302:
Checked link: http://www.plasticstorage.com/
Type of redirect: 302 Moved Temporarily
Redirected to: http://plasticstorage.com
Change the type of redirect to a 301 - this is done at the server level usually.
-
Unless I'm mistaken they appear to be identical duplicate pages.
-
little update:
In our GWT we have both the www version and the non www version.
On the non, we had set the preferred domain as the non www
On the WWW, we also had set the preferred domain as the non www
Does this mean I've done it correctly or not?
Thanks
-
I am looking into the 301/302 issue right now and i will report back...
Here are 3 supposed duplicate content pages:
<colgroup><col width="64"></colgroup>
|plasticstorage.com/catalog/product/view/id/5079/s/305b1/category/100/
plasticstorage.com/catalog/product/view/id/5079/s/305b1/category/101/
|
|
<colgroup><col width="64"></colgroup>
|Those pages when entered into your browser will appear identical... There is only one 305b1 item in the database. Somehow you can change the "100" or the "101" to "166666662" and you will still see that same "305b1" item.... Can you please help me figure out how or why and is this the "DUPLICATE CONTENT" or perhaps something else is being considered...
Thanks again
|
|
-
I'd still strongly recommend fixing the 301 redirect. The preference in the GWT tools is kinda like a soft preference.
Can you show 2 URSs that are deemed to be the same/duplicate?
-
We have set the canonical through Google to choose the non www version
Some examples of pages that are deemed duplicate is:
Page Title URL Other URLs Page Authority Linking Root Domains Ultra-Clear InSight Bin - Stacking 10x5x5 305B1 Akro-Mils
http://plasticstorage.com/305b1.html 50+ 15 1 Lids for Ultra-Clear InSight Bin - 305B1 - Akro-Mils 305B2
http://plasticstorage.com/305b2.html?related=1 50+ 1 0 Dividers for Ultra-Clear InSight Bin - 305B1 - Akro-Mils 405B1
http://plasticstorage.com/405b1.html?related=1 | 50+ | 1 |
| -
Hi there,
One issue is that your www.plasticstorage.com is a 302 redirect to plasticstorage.com. You should update this to a 301 redirect. This will be causing a few issues.
I think the SEOMoz tool shows content that it deems duplicate in the report. Can you post some examples of pages that are deemed duplicate?
Cheers.
-
I also had duplicate content. The cause was I have 2 similar domains that both led to the same page. Mine were (1) www.msperformanceonline.com & (2) w/o the " www." just msperformanceonline.com Look and see if this is why.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google Search Console issue: "This is how Googlebot saw the page" showing part of page being covered up
Hi everyone! Kind of a weird question here but I'll ask and see if anyone else has seen this: In Google Search Console when I do a fetch and render request for a specific site, the fetch and blocked resources all look A-OK. However, in the render, there's a large grey box (background of navigation) that covers up a significant amount of what is on the page. Attaching a screenshot. You can see the text start peeking out below (had to trim for confidentiality reasons). But behind that block of grey IS text. And text that apparently in the fetch part Googlebot does see and can crawl. My question: is this an issue? Should I be concerned about this visual look? Or no? Never have experienced an issue like that. I will say - trying to make a play at a featured snippet and can't seem to have Google display this page's information, despite it being the first result and the query showing a featured snippet of a result #4. I know that it isn't guaranteed for the #1 result but wonder if this has anything to do with why it isn't showing one. VmIqgFB.png
On-Page Optimization | | ChristianMKG0 -
How to organise subpages for good SEO content without duplicate text?
We are working on many subpages for our services. We have original content for each page however there are few text which we need to always duplicate like: Contact sales window, why to choose us window, supported files etc. What's the best way to do this so it's not consider as duplicated text. Should we redirected it or add it as a picture and always change name of the picture? Thank you Lukas
On-Page Optimization | | Lukas-ST0 -
Duplicate Content
I'm currently working on a site that sells appliances. Currently, there are thousands of "issues" with this site, many of them dealing with duplicate content. Now, the product pages can be viewed in "List" or "Grid" format. As Lists, they have very little in the way of content. My understanding is that the duplicate content arises from different URLs going to the same site. For instance, the site might have a different URL when told to display 9 items than when told to display 15. This could then be solved by inserting rel = canonical. Is there a way to take a site and get a list of all possible duplicates? This would be much easier than slogging through every iteration of the options and copying down the URLs. Also, is there anything I might be missing in terms of why there is duplicate content? Thank you.
On-Page Optimization | | David_Moceri0 -
Page content length...does it matter?
As I begin developing my website's content, does it matter how long or short the actual text found in the is? I heard someone say before "a minimum of 250 words", but is that true? If so, what is the maximum length I should use?
On-Page Optimization | | wlw20090 -
Should I worry about duplicate titles on pages where there is paginated content?
LivingThere.com is a real estate search site and many of our content pages are "search result" - ish in that a page often provides all the listings that are available and this may go on for multiple pages. For example, this is a primary page about a building: http://livingthere.com/building/31308-Cocoa-Exchange Because of the number of listings, the listings paginate to a second page: http://livingthere.com/building/31308-Cocoa-Exchange?MListings_page=2 Both pages have the same Page Title. Is this a concern? If so is there a "best practice" for giving paginated content different titles? Thanks! Nate
On-Page Optimization | | nate1230 -
Duplicate Page Content Question
This article was published on fastcompany.com on March 19th. http://www.fastcompany.com/magazine/164/designing-facebook It did not receive much traffic, so it was re-posted on Co.Design today (March 27th) where it has received significantly more traffic. http://www.fastcodesign.com/1669366/facebook-agrees-the-secret-to-its-future-success-is-design My question is if google will dock us for reprinting/reusing content on another site (even if it is a sister site within the same company). If they do frown on that, is there a proper way to attribute the content to the source material/site (fastcompany.com)?
On-Page Optimization | | DanAsadorian0 -
DUPLICATE PAGE VIA ONPAGE SEARCH
Hi, we have a fashion shop and we try to keep the page title as unique as possible. However google also indexes the search so the say search for brand gucci - type - dress - size 8 - then it indexes it with the homepage title rather then what the searched for. Should I block google from indexing search variation (because there could be thousand of variations) or is there a module that would use the search as a page title so gucci dress 8 for example. Many thanks for your time.
On-Page Optimization | | reallyitsme0 -
Duplicate content issue with dynamically generated url
Hi, For those who have followed my previous question, I have a similar one regarding dynamically generated urls. From this page http://www.selectcaribbean.com/listing.html the user can make a selection according to various criteria. 6 results are presented and then the user can go to the next page. I know I should probably rewrite url's such as these: http://www.selectcaribbean.com/listing.html?pageNo=1&selType=&selCity=&selPrice=&selBeds=&selTrad=&selMod=&selOcean= but since all the results presented are basically generated on the fly for the convenience of the user, I am afraid google my consider this as an attempt to generate more pages as there are pages for each individual listing. What is my solution for this? Nofollow these pages? Block them thru robots txt?
On-Page Optimization | | multilang0