SEOmoz suddenly reporting duplicate content with no changes???
-
I am told the crawler has been updated and wanted to know if anyone else is seeing the same thing I am.
SEOmoz reports show many months of no duplicate content problems. As of last week though, I get a little over a thousand pages reported as dupe content errors.
Checking these pages I find there is similar content (hasn't changed) with keywords that are definitely different. Many of these pages rank well in Google, but SEOmoz is calling them out as duplicate content. Is SEOmoz attempting to closely imitate Google's perspective in this matter and therefore telling me that I need to seriously change the similar content?
Anyone else seeing something like this?
-
Hi
We see the extreme raise in the duplicate content of our site too. If the sensitivity is adapted, will these graphs come down again?
What is your opinion on how Google sees a webshop with lot's of products and filter options? Our site www.dmlights.com/massive for example can have a lot of filtering but we try to counter this in Webmaster Tools with the URL parameters.
Do you suggest to adapt this for good seo?
Wondering about your opinions. Thanks.
-
Hey Scott,
Again, we're sorry about the odd jump in duplicate content errors!
We just launched a new crawler and it is being extremely sensitive to duplicate content. As of now we are picking up duplicate pages on your domain via:
https clones of URLs
Some pages have a “/” trailing after the URL and some don’t
We are also ignoring some rel=canonical directives
This is an issue that other users are seeing with their crawls. Our engineers have made some changes to the crawler to scale back the sensitivity to these issues on the crawler and you should be seeing the changes within a week or two.
We're really sorry for the confusion.
Best of Luck,
Chiaryn
-
Two good suggestions so far, and both I had checked. Thanks KJ Rogers and Ryan Kent.
This is starting to look like it boils down to how much the new SEOmoz crawler sees content in the same way that Google does.
We did not make any site-wide changes and the URLs identified as duplicate in the report are valid URLs that actually hold similar content (keywords and so forth were changed for each version of a slightly different product through an Excel Concatenate construct to build the content). We have actually seen these pages climb in rank over the months since the content was added.
So, like I said, the sudden identification of these as duplicate by the moz crawler is suspicious to me. Not sure it sees things the way Google does.
-
Without examining your site and the pages involved it is not possible for me to share feedback.
Is it possible you made any recent site wide changes? Changes to your header, navigation, footer or sidebar could have pushed you passed a certain threshhold of duplicate content which triggered a flag.
-
I got the same thing last week. I later found out that mine, using dynamic content on the same page, had speical characters in the url which was taking crawlers to an error page. The error page was showing a list of pages with the url's capitalized. I was able to fix some of them, but it scared the heck out of me.
I had to run a crawl test from SEOMoz to filter out what was going on. Perhaps you have something similar?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is good for SEO update blog post dates after update post content
Hello I am updating some posts of my Blog, adding new and fresh content and rewriting some of the existing. After doing that I am thinking to update de post publishing so that I appears on front page of the blog and user can read ir again. But I don't know if it is good for google to change the publishing date of the post that he had indexed 5 years ago. Also I don't know if google will read it again if it is old and see the new changes in order to improve it in search results
Algorithm Updates | | maestrosonrisas0 -
SEO Myth-Busters -- Isn't there a "duplicate content" penalty by another name here?
Where is that guy with the mustache in the funny hat and the geek when you truly need them? So SEL (SearchEngineLand) said recently that there's no such thing as "duplicate content" penalties. http://searchengineland.com/myth-duplicate-content-penalty-259657 by the way, I'd love to get Rand or Eric or others Mozzers aka TAGFEE'ers to weigh in here on this if possible. The reason for this question is to double check a possible 'duplicate content" type penalty (possibly by another name?) that might accrue in the following situation. 1 - Assume a domain has a 30 Domain Authority (per OSE) 2 - The site on the current domain has about 100 pages - all hand coded. Things do very well in SEO because we designed it to do so.... The site is about 6 years in the current incarnation, with a very simple e-commerce cart (again basically hand coded). I will not name the site for obvious reasons. 3 - Business is good. We're upgrading to a new CMS. (hooray!) In doing so we are implementing categories and faceted search (with plans to try to keep the site to under 100 new "pages" using a combination of rel canonical and noindex. I will also not name the CMS for obvious reasons. In simple terms, as the site is built out and launched in the next 60 - 90 days, and assume we have 500 products and 100 categories, that yields at least 50,000 pages - and with other aspects of the faceted search, it could create easily 10X that many pages. 4 - in ScreamingFrog tests of the DEV site, it is quite evident that there are many tens of thousands of unique urls that are basically the textbook illustration of a duplicate content nightmare. ScreamingFrog has also been known to crash while spidering, and we've discovered thousands of URLS of live sites using the same CMS. There is no question that spiders are somehow triggering some sort of infinite page generation - and we can see that both on our DEV site as well as out in the wild (in Google's Supplemental Index). 5 - Since there is no "duplicate content penalty" and there never was - are there other risks here that are caused by infinite page generation?? Like burning up a theoretical "crawl budget" or having the bots miss pages or other negative consequences? 6 - Is it also possible that bumping a site that ranks well for 100 pages up to 10,000 pages or more might very well have a linkuice penalty as a result of all this (honest but inadvertent) duplicate content? In otherwords, is inbound linkjuice and ranking power essentially divided by the number of pages on a site? Sure, it may be some what mediated by internal page linkjuice, but what's are the actual big-dog issues here? So has SEL's "duplicate content myth" truly been myth-busted in this particular situation? ??? Thanks a million! 200.gif#12
Algorithm Updates | | seo_plus0 -
New Website Old Domain - Still Poor Rankings after 1 Year - Tagging & Content the culprit?
I've run a live wedding band in Boston for almost 30 years, that used to rank very well in organic search. I was hit by the Panda Updates August of 2014, and rankings literally vanished. I hired an SEO company to rectify the situation and create a new WordPress website -which launched January 15, 2015. Kept my old domain: www.shineband.com Rankings remained pretty much non-existent. I was then told that 10% of my links were bad. After lots of grunt work, I sent in a disavow request in early June via Google Wemaster Tools. It's now mid October, rankings have remained pretty much non-existent. Without much experience, I got Moz Pro to help take control of my own SEO and help identify some problems (over 60 pages of medium priority issues: title tag character length and meta description). Also some helpful reports by www.siteliner.com and www.feinternational.com both mentioned a Duplicate Content issue. I had old blog posts from a different domain (now 301 redirecting to the main site) migrated to my new website's internal blog, http://www.shineband.com/best-boston-wedding-band-blog/ as suggested by the SEO company I hired. It appears that by doing that -the the older blog posts show as pages in the back end of WordPress with the poor meta and tile issues AS WELL AS probably creating a primary reason for duplicate content issues (with links back to the site). Could this most likely be viewed as spamming or (unofficial) SEO penalty? As SEO companies far and wide daily try to persuade me to hire them to fix my ranking -can't say I trust much. My plan: put most of the old blog posts into the Trash, via WordPress -rather than try and optimize each page (over 60) adjusting tagging, titles and duplicate content. Nobody really reads a quick post from 2009... I believe this could be beneficial and that those pages are more hurtful than helpful. Is that a bad idea, not knowing if those pages carry much juice? Realize my domain authority not great. No grand expectations, but is this a good move? What would be my next step afterwards, some kind of resubmitting of the site, then? This has been painful, business has fallen, can't through more dough at this. THANK YOU!
Algorithm Updates | | Shineband1 -
Google not crawling click to expand content - suggestions?
It seems like Google confirmed this week in a G+ hangout that content in click to expand content e.g. 'read more' dropdown and tabbed content scenarios will be discounted. The suggestion was if you have content it needs to be visible on page load. Here's more on it https://www.seroundtable.com/google-index-click-to-expand-19449.html and the actual hangout, circa 11 mins in https://plus.google.com/events/cjcubhctfdmckph433d00cro9as. From a UX and usability point of view having a lot of content that was otherwise tabbed or in click to expand divs can be terrible, especially on mobile. Does anyone have workable solutions or can think of examples of really great landing pages (i'm mostly thinking ecommerce) that also has a lot of visible content? Thanks Andy
Algorithm Updates | | AndyMacLean0 -
Content on Wordpress blog inside the main website for SEO
Hi, We have our main website and our blog on blog.practo.com. Now what I see is that we wish to write in content to grow our seo keywords and links. Should we put the blog as www.practo.com/blog and then begin writing all the content or we should put the wordpress blog as www.practo.com/(wordpress blog here) and then begin writing the content. For best practices I suppose we should have content lined up as www.sitename.com/category/article name etc or www.sitename.com/article name etc - am I correct? Our main site consists of few html pages and then we have our software on a different sub domain. What are the best ways to publish content and get it crawled at a faster rate for growth? I would also wish to understand how to measure the number of growth in % to our content we are writing. Only via google analytics or some other tool? Say I wish to see the growth of 10 articles from month of may and compare it to the month of april or march 2012. So what tools could I use to see if we are progressing or not? Thanks
Algorithm Updates | | shanky10 -
Does this mean that exact keyword phrase anchor text is not the dominating ranking factor anymore for serps?http://insidesearch.blogspot.com/2011/11/ten-recent-algorithm-changes.html
Does this mean that exact keyword phrase anchor text is not the dominating ranking factor anymore for serps? http://insidesearch.blogspot.com/2011/11/ten-recent-algorithm-changes.html If so what is the new most important factor?
Algorithm Updates | | AndrewSEO0 -
New Algorithm changes
the real time news feed style in which we all use the internet these days. How do you think this new change is going to effect things ? “This is the result of them saying we need to find a way to more effectively get fresh content up,” said Danny Sullivan, editor of Search Engine Land and an industry expert. “It does help with the issue of people thinking, ‘Wow, if I need to find out about something breaking, I’ll go to Facebook or Twitter for that.’ ” Is google reacting to a massive loss of traffic volume from Facebook and Twitter ? I also ask the question would Facebook benifit from some form of built in search engine or would this never happen ??
Algorithm Updates | | onlinemediadirect0 -
High bounce rates from content articles influencing our rankings for rest of site
We have a large content article section on our e-commerce site that receives a lot of visits but also have very high bounce rates. We are wondering if this is hurting the rest of our site's rankings. **When I say bounce rates I mean what ever metrics Google is using to determine quality content (specifically after the Panda update). ** We are trying to determine if having the content articles on our domain hurts us. We only have the content articles for link building.
Algorithm Updates | | seozachz0