Tags creating duplicated content issue?
-
Hello i believe a lot of us use tags in our blogs as a way to categorize content and make it easy searchable but this usually (at lease in my case) cause duplicate content creation.
For example, if one article has 2 tags like "SEO" & "Marketing", then this article will be visible and listed in 2 urls inside the blog like this
In case of a blog with 300+ posts and dozens of different tags this is creating a huge issue.
My question is 1. Is this really bad? 2. If yes how to fix it without removing tags?
-
I have different meta content since a long time still showing as a duplicate and on just looking at the body content it is identical. Is there any quick way I can manually add something to the robots file to take the duplicates away? Canonical is not working for me as it just points to the same url - not the MAIN one you want. So there is nothing as good as yoast for joomla, they should make that and make a lot of money! OOTB joomla is poor at seo if you dont know how to make menus in joomla your site can have massive issues. Without a tool like MOZ you may never know why your quality content can't rank - gee thanks joomla
-
Hi
Ahhhh... gotcha thought it was wordpress
Your best bet is to have a unique description generated in Joomla for each tag archive. Robots.txt won't necessarily remove the URLs from Google. If you want to deindex them, you need to use meta robots tag.
Anyhow, hope that got you in the right direction!
-Dan
-
Dear Dan,
Thank you so much for spending time on our issue and on the advice. Im looking forward to read your article.
Unfortunetly our blog for technical purposes is not in Wordpress but in Joomla, so i will look for a similar solution there. The desperate solution i guess is to disallow tag urls in robots.txt. But i would try avoid that. On the other hand, since i also use categories to index the content then i assume this will not generate any issue of hiding content.
-
Hey Guys
Again, whether full posts or excerpts are being shown for tag archives, is important (I would vote on excerpts) but see my answer above. The tag archives all have the same description. That's where Moz is likely getting the duplicate errors from - and not likely because of tag pages being similar to post or category pages.
The quick fix on this is to use an SEO plugin like Yoast and create a description template for the tag archives.
But BEST case scenario in a perfect setup, would be have tags totally unique from categories, and not index tag archives at all.
Canonicals should only be used sparingly and when no other measure can be taken.
It also seems this is not the best theme, so there are other issues at play as well, too many to go through in just a Q&A format.
-Dan
-
Hi
Just want to add two cents to this... a canonical should really be the last resort if it can't be resolved with robots meta, url structure, or content.
The issue here is that Moz is bringing back duplicate content errors because the tags all have the same description. This can be fixed (as noted in my full answer) by creating a description template for tag archives with a plugin like Yoast SEO.
The canonical may not resolve anything because the tag pages at best shouldn't be indexed to begin with - and if they are indexed, the descriptions should be unique.
-Dan
Edit - just realized they are using Joomla. The same can apply, but I'm not as familiar with Joomla, so if there's a way to create descriptions for the tags with Joomla that's the best bet still.
-
Hi!
Just need to clear things up here, sorry I'm a little late to responding!
1. Quick Fix - Create a description template for tag archives
You're getting duplicate errors because your tag archives all have the same meta description. Use an SEO plugin like Yoast SEO for wordpress something for Joomla and create a template for your tag descriptions. This will give each tag archive a unique description and eliminate the duplicate errors.
2. Long Term Fix - Root of The Problem
The real ROOT of the issue, is a combination of maybe a poor theme, no SEO plugin (that I can see) and tag pages being used incorrectly.
-
Tags should be completely different than any categories
-
And as standard practice I NOINDEX tags. Because there content is so similar to other pages, and it also may not be the best user experience. There may be exceptions to this but its a general rule I follow.
Now, with that said, don't just go deindexing your tag archives.
Tomorrow (May 8th 2012), I have an extensive article going up on the Moz blog about WordPress and duplicate content. I suggest reading that article to get a good understanding of how all the elements work. And perhaps in the long term you can work towards a more robust WordPress setup. But for now, no harm done the way it is.
Hope that helps!
-Dan
Edit - Realized they are using Joomla. The same concepts apply, but with a technical implementation that works with Joomla (which I am not as familiar with).
-
-
I think you should be good leaving it alone, then.
You could put rel=canonical on the post page only (don't put it on the tags or category pages) but that might be more trouble than it's worth, depending on the restrictions imposed by the CMS.
-
I don't believe the actual tag pages are the issue here. It's the fact that the same page can be accessed by 3 different url's because of the tags it's under. Canonical links will take care of this.
-
I am not sure if it's possible with the publishing system you are working with, but there are CMS systems on the market who have solved this issue.
They have done the following approach:
Create your main Article, Blog etc., tag them with your keywords and on your keyword page show the Article content as a teaser with a ''Read More" link to the full content page.
This is not considered as duplicate content!
Hope this helps!
-
Hi Pantelis,
I think that whether or not this is a problem, and how it should be fixed, depends on how your blog is set up.
The guide Justin mentioned is a good resource. Before you jump in, I think you should consider these questions:
When you go to domain.com/blog/seo etc. are the posts excerpted, or are full posts being displayed?
When someone clicks on the title of a blog post having found it under a tag (e.g. going to domain.com/blog/marketing and clicking on one of the posts) what URL is being displayed for the individual post?
e.g. is it domain.com/blog/seo/great-post-1 or is it domain.com/blog/great-post-1 ?
What really matters for duplicate content and canonicalization is whether the URL for the individual blog post is unique.
If the blog post has one unique URL, no matter how you get to it, and if the tags pages are displaying excerpts, then the only place you should be using rel=canonical is on the blog post itself. I think putting rel=canonical on a tags page that's only displaying titles and excerpts is asking for trouble. I don't like the idea of the search engines potentially thinking that your tag page, which has partials of many posts, is the original source.
If you're displaying full blog posts on the tags pages, then the solution is probably to switch it to excerpts and canonicalize only the individual blog posts.
Reference the SEOmoz blog: The SEOmoz.org/blog page doesn't use rel=canonical, and only displays excerpts, while seomoz.org/blog/post-title uses rel=canonical and displays the full post.
-
Its not really bad, but there is every chance it will affect your rankings as google will not know which page is dominant and in turn will not know which version it should show to searchers
the best method of resolving the issue is to use the rel=canonical tag as this allows you to tell google which page is the dominant version
see article here for more details:
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Exclude sorting options using nofollow to reduce duplicate content
I'm getting reports of duplicate content for pages that have different sorting options applied, e.g: /trips/dest/africa-and-middle-east/
On-Page Optimization | | benbrowning
/trips/dest/africa-and-middle-east/?sort=title&direction=asc&page=1
/trips/dest/africa-and-middle-east/?sort=title&direction=des&page=1 I have the added complication of having pagination combined with these sorting options. I also don't have the option of a view all page. I'm considering adding rel="nofollow" to the sorting controls so they are just taken out of the equation, then using rel="next" and rel="prev" to handle the pagination as per Google recommendations(using the default sorting options). Has anyone tried this approach, or have an opinion on whether it would work?0 -
Duplicate content penalty
when moz crawls my site they say I have 2x the pages that I really have & they say I am being penalized for duplicate content. I know years ago I had my old domain resolve over to my new domain. Its the only thing that makes sense as to the duplicate content but would search engines really penalize me for that? It is technically only on 1 site. My business took a significant sales hit starting early July 2013, I know google did and algorithm update that did have SEO aspects. I need to resolve the problem so I can stay in business
On-Page Optimization | | cheaptubes0 -
Duplicate Content
I'm currently working on a site that sells appliances. Currently, there are thousands of "issues" with this site, many of them dealing with duplicate content. Now, the product pages can be viewed in "List" or "Grid" format. As Lists, they have very little in the way of content. My understanding is that the duplicate content arises from different URLs going to the same site. For instance, the site might have a different URL when told to display 9 items than when told to display 15. This could then be solved by inserting rel = canonical. Is there a way to take a site and get a list of all possible duplicates? This would be much easier than slogging through every iteration of the options and copying down the URLs. Also, is there anything I might be missing in terms of why there is duplicate content? Thank you.
On-Page Optimization | | David_Moceri0 -
How should I rephrase these pages to avoid Phrase duplication within Title Tags
How should I rephrase these pages to avoid Phrase duplication within Title Tags Duplicate Page Title Page1-http://organicfruitbasketsflorist.com/index-2.html Page2- http://organicfruitbasketsflorist.com/Fruit_Baskets_Organic_Fruit_Baskets_New_York_NY.html Page3- http://organicfruitbasketsflorist.com/Fruit_Baskets_Edible_Fruit_Baskets_New_York_NY.html Page4organicfruitbasketsflorist.com/Fruit_Baskets_Business_Fruit_Baskets_New_York_NY.html Page5-http://organicfruitbasketsflorist.com/Fruit_Baskets_Fresh_Flowers_Delivered_New_York_NY.html Page6-http://organicfruitbasketsflorist.com/Coupons.htmlAmi
On-Page Optimization | | amydiamond0 -
Exponentially Increasing Duplicate Content On Blogs
Most of the clients that I pick up are either new to SEO best practices, or have worked with sketchy SEO providers in the past, who did little more than build spammy links. Most of them have deployed little if any on-site SEO best practices, and early on I spend a lot of time fixing canonical and duplicate content issues alla 301 redirects. Using SEOMOZ, however, I see a lot of duplicate content issues with blogs that live on the sites I work on. With every new blog article we publish, more duplicate content builds up. I feel like duplicate content on blogs grows exponentially, because every time you write a blog article, it exists provisionally on the blog homepage, the article link, a category page, maybe a tag page, and an author page. I have a two-part question: Is duplicate content like this a problem for a blog -- and for the website that the blog lives on? Are search engines able to parse out that this isn't really duplicate content? If it is a problem, how would you go about solving it? Thanks in advance!
On-Page Optimization | | RCNOnlineMarketing0 -
Duplicate content
crawler shows following links as duplicate http://www.mysite.com http://mysite.com http://www.mysite.com/ http://mysite.com. http://mysite.com/index.html How can i solve this issue?
On-Page Optimization | | bhanu22170 -
Duplicate Content Warning
Hi Mozers, I have a question about the duplicate content warnings I am recieving for some of my pages. I noticed that the below pattern of URLs are being flagged as duplicate content. I understand that these are seen as two different pages but I would like to know if this has an negative impact on my SEO? Why is this happening? How do I stop it from happening? http://www.XXXX.com/product1234.html?sef_rewrite=1 http://www.XXXX.com/product1234.html Thanks in advance!
On-Page Optimization | | mozmonkey0 -
Percentage of duplicate content allowable
Can you have ANY duplicate content on a page or will the page get penalized by Google? For example if you used a paragraph of Wikipedia content for a definition/description of a medical term, but wrapped it in unique content is that OK or will that land you in the Google / Panda doghouse? If some level of duplicate content is allowable, is there a general rule of thumb ratio unique-to-duplicate content? thanks!
On-Page Optimization | | sportstvjobs0