WordPress Duplicate Content Issues
-
Everyone knows that WordPress has some duplicate content issues with tags, archive pages, category pages etc...
My question is, how do you handle these issues?
Is the smart strategy to use robots meta and add no follow/ no index category pages, archive pages tag pages etc?
By doing this are you missing out on the additional internal links to your important pages from you category pages and tag pages?
I hope this makes sense.
Regards,
Bill
-
Hey Bill
I like to start with this standard setup (image/chart from my wordpress post on moz);
Pages, Posts, Categories - Index
Tags, Dated Archives, Subpages, Author Archives - noindex
You can check out the full post - I will be updating the Yoast Screenshots very soon!
-Dan
-
Thanks for article,
Now 2 years ahead, are there any important updates for preventing duplicate content/titles?
-
Most of the Plugins for wordpress use canonical urls.
-
Unless I'm missing something here, wouldn't it be easier to set the canonical tag for the main post? There are also plugins like SEO Ultimate that handle this automatically.
-
I posted this article I wrote the other day for someone asking a similar question.
With the Yoast SEO Plugin I no-index everything except Categories. You can see how I set mine up under section 3. Indexation.
Here is the original question that Sha submitted:
http://www.seomoz.org/q/what-is-with-wordpress-dupe-issues -
Bill-
There are several SEO plugs available for WP that will handle these issues. Yes, you are right that adding "noindex" will be beneficial on tag, category, and archive pages. The idea here is avoiding duplicate content issues. BTW, check out: Yoast SEO for Wordpress.
Here is how the values for the robots meta tag work:
- noindex will keep a page from being crawled
- nofollow will prevent a page's links from being followed
I agree with noindex'ing these pages; though I would argue that a nofollow is still worth leaving out. If these pages have any juice you want to allow this to flow to the other links on the page.
-
The WP on my blog is set up as follows (this is a blog that gets between four and ten short posts per day - about two to four sentences, each post linking to an article or other content on a topic-related website)
Homepage: Full text of the most recent 25 posts are displayed. Pagination pages are not indexed (blocked by robots.txt).
Post Pages: Full text is displayed and the title plus a few words of 20 related posts are displayed.
Category Pages: I have over 100 categories and each post is placed into at least two categories (one by location and one by topic). Some posts go into three or four categoreis - sometimes more. Each category page displays the full text of the most recent 25 posts. Categories do not have pagination pages (blocked by robots.txt).
All of the above pages are fully indexed and a long list of category pages appears in the left-side navigation. I don't use tag pages or archive pages. There is a lot of dupe content in this system but so far I am lucky that it does not cause a problem. The category pages pull a lot of organic search traffic.
In January of each year I delete all of the posts that are over a year old. Before doing that I identify those that are pulling reasonable traffic and either redirect them to a permanent page about same topic, write an article about that topic and redirect, or recycle that post. All the rest are redirected to the homepage of the blog.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Recurring events and duplicate content
Does anyone have tips on how to work in an event system to avoid duplicate content in regards to recurring events? How do I best utilize on-page optimization?
Technical SEO | | megan.helmer0 -
Duplicate Content - Different URLs and Content on each
Seeing a lot of duplicate content instances of seemingly unrelated pages. For instance, http://www.rushimprint.com/custom-bluetooth-speakers.html?from=topnav3 is being tracked as a duplicate of http://www.rushimprint.com/custom-planners-diaries.html?resultsperpg=viewall. Does anyone else see this issue? Is there a solution anyone is aware of?
Technical SEO | | ClaytonKendall0 -
Minimising the effects of duplicate content
Hello, We realised that one of our clients, copied a large part of content from our website to his. The normal reaction would be to send a cease and desist letter. Nevertheless this would probably mean loosing a good client. The client dumped the text of several articles (for example:
Technical SEO | | Lvet
http://www.velascolawyers.com/en/property-law/136-the-ley-de-costas-coastal-law.html ) Into the same page:
http://www.freundlinger-partners.com/en/home/faqs-property-law/ I convinced the client to place our authorship tags on this page, but I am wondering if this is enough. What do you think? Cheers
Luca0 -
Looking for a technical solution for duplicate content
Hello, Are there any technical solutions to duplicate content similar to the nofollow tag? A tag which can indicate to Google that we know that this is duplicate content but we want it there because it makes sense to the user. Thank you.
Technical SEO | | FusionMediaLimited0 -
Duplicate Content
SEOmoz is reporting duplicate content for 2000 of my pages. For example, these are reported as duplicate content: http://curatorseye.com/Name=“Holster-Atlas”---Used-by-British-Officers-in-the-Revolution&Item=4158
Technical SEO | | jplill
http://curatorseye.com/Name=âHolster-Atlasâ---Used-by-British-Officers-in-the-Revolution&Item=4158 The actual link on the site is http://www.curatorseye.com/Name=“Holster-Atlas”---Used-by-British-Officers-in-the-Revolution&Item=4158 Any insight on how to fix this? I'm not sure where the second version of the URL is coming from. Thanks,
Janet0 -
Rel=canonical overkill on duplicate content?
Our site has many different health centers - many of which contain duplicate content since there is topic crossover between health centers. I am using rel canonical to deal with this. My question is this: Is there a tipping point for duplicate content where Google might begin to penalize a site even if it has the rel canonical tags in place on cloned content? As an extreme example, a site could have 10 pieces of original content, but could then clone and organize this content in 5 different directories across the site each with a new url. This would ultimately result in the site having more "cloned" content than original content. Is this at all problematic even if the rel canonical is in place on all cloned content? Thanks in advance for any replies. Eric
Technical SEO | | Eric_Lifescript0 -
I'm getting a Duplicate Content error in my Pro Dashboard for 2 versions of my Homepage. What is the best way to handle this issue?
Hi SEOMoz,I am trying to fix the final issues in my site crawl. One that confuses me is this canonical homepage URL fix. It says I have duplicate content on the following pages:http://www.accupos.com/http://www.accupos.com/index.phpWhat would be the best way to fix this problem? (...the first URL has a higher page authority by 10 points and 100+ more inbound links).Respectfully Yours,Derek M.
Technical SEO | | DerekM880 -
Are recipes excluded from duplicate content?
Does anyone know how recipes are treated by search engines? For example, I know press releases are expected to have lots of duplicates out there so they aren't penalized. Does anyone know if recipes are treated the same way. For example, if you Google "three cheese beef pasta shells" you get the first two results with identical content.
Technical SEO | | RiseSEO0