Duplicate content - news archive
-
Most of them are due to news items having more than 1 category – which is pretty normal.Also /us/blog, /uk/blog and /ca/blog are effectively the same page.None of them are actually duplicate content – just alternate URLs for the same pagehttp://www.fdmgroup.com/category/news/
-
From developer: "Looking into this, we need to have /uk/blog, /us/blog and /ca/blog in order for them to appear on the menus – we could put a noindex meta tag on the us and ca pages to avoid duplicates?"
Or do you recommend href lang tag? Thanks.
-
From developer: "Looking into this, we need to have /uk/blog, /us/blog and /ca/blog in order for them to appear on the menus – we could put a noindex meta tag on the us and ca pages to avoid duplicates?"
-
Hi Christopher,
Google has definitely become a lot better in recent years at identifying this sort of duplication and dealing with it (largely because this has to be one of the most common accidental / non-malicious duplication causes - categories on blogs and news sites). That said, cleaning it up is for the best. I have been meaning to clean up a blog belonging to a family member in this manner for months (years...) because the version of each piece of content Google has chosen is wrong. Pages marked up with dates, e.g. example.com/2014/01 kept ranking better than the original posts in that date range. So even when Google makes the decision for you, it won't necessarily make the right one. You're risking visitors coming to a page they didn't expect, or a page that doesn't answer their query as succinctly as the "best" version would have, and if you are in e-commerce of any sort or focusing on conversions, this can make a big difference to how optimised your traffic's on-site experience is.
Where you'll be "penalised" for duplicate content, especially by Panda, is as you cite above: when the duplication looks like it has been done for spam purposes. This has happened accidentally to people when their content management systems have gone mad with infinite duplication, but it likely won't happen with simple blog categories.
In short, Google sees this sort of duplication all day, every day and will choose its favourite version to rank. However, if you can guide its choice, you're in control of what your visitors see.
You mention country-based categories in your original question. If internationalisation and duplicate content are a concern, you might want to check out the href lang tag (also called rel="alternate" tag - it gets called either by the community). Could be useful if you're publishing the same thing in different countries.
-
From my developer:"Doing a bit of research Google have explicitly stated that that don’t penalise duplicate content unless it appears to be deliberately deceptive. The only issue is which version appears in the search results.Duplicate content on a site is not grounds for action on that site unless it appears that the intent of the duplicate content is to be deceptive and manipulate search engine results. If your site suffers from duplicate content issues, and you don't follow the advice listed above, we do a good job of choosing a version of the content to show in our search results.https://support.google.com/webmasters/answer/66359?hl=enMatt Cutts, Google’s head of search spam, posted a video today about duplicate content and the repercussions of it within Google’s search results.Matt said that somewhere between 25% to 30% of the content on the web is duplicative. Of all the web pages and content across the internet, over one-quarter of it is repetitive or duplicative.But Cutts says you don’t have to worry about it. Google doesn’t treat duplicate content as spam. It is true that Google only wants to show one of those pages in their search results, which may feel like a penalty if your content is not chosen — but it is not.Google takes all the duplicates and groups them into a cluster. Then Google will show the best of the results in that cluster.Matt Cutts did say Google does reserve the right to penalize a site that is excessively duplicating content, in a manipulative manner. But overall, duplicate content is normal and not spam.http://searchengineland.com/googles-matt-cutts-25-30-of-the-webs-content-is-duplicate-content-thats-okay-180063http://searchengineland.com/googles-matt-cutts-duplicate-content-wont-hurt-you-unless-it-is-spammy-167459Cheers"
-
I'm afraid your blog pages are in fact duplicate content, in Google's eyes anyway.
The /us/blog, /uk/blog and /ca/blog examples are all separate URLs that you are asking Google to index (separate canonical tags for each and no robots instructions that I can see). Google is going to look at these and any blog posts within them as separate pages. Once it realises they all have the same content, it will likely result in a Panda algorithmic penalty.
The risk here is that this penalty might affect your entire domain, rather than the offending pages. I really don't see that as a risk worth taking. Therefore, I strongly advise to remove the separate versions of the blogs and consolidate into one blog, with redirections of the local blogs to the new ones. Failing that, choose one version and instruct Google not to index other versions of the page by using a meta robots tag in your header, or in the robots.txt file.
I also advise that you noindex the category page to be sure that its content isn't being seen as duplicate either. More info on how to do that can be found in the Moz Robots Guide.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
"Duplicate without user-selected canonical” - impact to SERPs
Hello, we are facing some issues on our project and we would like to get some advice. Scenario
International SEO | | Alex_Pisa
We run several websites (www.brandName.com, www.brandName.be, www.brandName.ch, etc..) all in French language . All sites have nearly the same content & structure, only minor text (some headings and phone numbers due to different countries are different). There are many good quality pages, but again they are the same over all domains. Goal
We want local domains (be, ch, fr, etc.) to appear in SERPs and also comply with Google policy of local language variants and/or canonical links. Current solution
Currently we don’t use canonicals, instead we use rel="alternate" hreflang="x-default": <link rel="alternate" hreflang="fr-BE" href="https://www.brandName.be/" /> <link rel="alternate" hreflang="fr-CA" href="https://www.brandName.ca/" /> <link rel="alternate" hreflang="fr-CH" href="https://www.brandName.ch/" /> <link rel="alternate" hreflang="fr-FR" href="https://www.brandName.fr/" /> <link rel="alternate" hreflang="fr-LU" href="https://www.brandName.lu/" /> <link rel="alternate" hreflang="x-default" href="https://www.brandName.com/" /> Issue
After Googlebot crawled the websites we see lot of “Duplicate without user-selected canonical” in Coverage/Excluded report (Google Search Console) for most domains. When we inspect some of those URLs we can see Google has decided that canonical URL points to (example): User-declared canonical: None
Google-selected canonical: …same page, but on a different domain Strange is that even those URLs are on Google and can be found in SERPs. Obviously Google doesn’t know what to make of it. We noticed many websites in the same scenario use a self-referencing approach which is not really “kosher” - we are afraid if we use the same approach we can get penalized by Google. Question: What do you suggest to fix the “Duplicate without user-selected canonical” in our scenario? Any suggestions/ideas appreciated, thanks. Regards.0 -
Hreflang tags and canonical tags - might be causing indexing and duplicate content issues
Hi, Let's say I have a site located at https://www.example.com, and also have subdirectories setup for different languages. For example: https://www.example.com/es_ES/ https://www.example.com/fr_FR/ https://www.example.com/it_IT/ My Spanish version currently has the following hreflang tags and canonical tag implemented: My robots.txt file is blocking all of my language subdirectories. For example: User-agent:* Disallow: /es_ES/ Disallow: /fr_FR/ Disallow: /it_IT/ This setup doesn't seem right. I don't think I should be blocking the language-specific subdirectories via robots.txt What are your thoughts? Does my hreflang tag and canonical tag implementation look correct to you? Should I be doing this differently? I would greatly appreciate your feedback and/or suggestions.
International SEO | | Avid_Demand0 -
Do I have duplicate content issues to be worried about?
Hey guys, We built a website http://www.cylon.com/ targeting different regions but with the same English langauage (Ireland, England and America). The content for the most part is the same set up on 3 different subfolders. http://www.cylon.com/ - Targeting United States in WMT http://www.cylon.com/ie - Targeting Ireland in WMT http://www.cylon.com/uk - Targeting UK in WMT Do I have duplicate content issues to be worried about? If so, how do I get around this issue? Also is there anyway of finding out if Google have in some way penalised these pages for having the same content on other pages trageting different Countries? I have not received any messages from Google in WMT saying there is duplicate so I'm not sure if this is an issue. Thanks Rob
International SEO | | daracreative0 -
Delivering different content according to country
Hey, I have a question regarding different content according to country (IP)-
International SEO | | Kung_fu_Panda
We planing to serve mobile users using dynamic HTML serving (on the same url)
Is it possible to serve different content for different devices + different IPs (for example different content for a user from US android and someone from UK android ) thanks!0 -
Should I be deindexing pages with thin or weak content?
If I have pages that rank product categories by alphabetical order should I deindex those pages? Keeping in mind the pages do not have any content apart from product titles? For example: www.url.com/albums/a/ www.url.com/albums/b/ If I deindexed these pages would I lose any authority passed through internal linking?
International SEO | | Jonathan_Hatton0 -
International hreflang - will this handle duplicate content?
The title says it all - if i have duplicate content on my US and UK website, will adding the hreflang tag help google figure out that they are duplicate for a reason and avoid any penalties?
International SEO | | ALLee1 -
Looking for content writers for multi-language SEO
Hi All, I'm currently doing a lot of work for a UK client who has multiple sites outside the UK (all part of the same business). We're currently discussing the option of us handling all of his SEO for his German, French, Spanish and Italian sites too, but we only have access to one person in the office who can speak French and Spanish. They're currently booked up on other jobs that we can't really move them off, so I'm looking for options of outsourcing some of the content writing. My question is, does anyone know of any high quality content writing services that have writers available to write for the countries languages above? We're going to focus initially on their on-site strategy and building up their high quality content. At the moment, they don't have much relevant content on their website, so we're going to initially look at this. Moving forward, we'll be looking at their off-site strategy and trying to find areas to submit high quality articles, look at guest blogging and PR opportunities. Any tips anyone has on this side (in terms of outsourcing to native speakers) would be quite useful too! Many thanks,
International SEO | | PinpointDesigns
Lewis0 -
Is duplicate content a concern across multiple CCTLDs?
Looking for experienced feedback on a new client challenge. Multiple pages of content in the English language are replicated across multiple CCTLDs in addition to the .com address we're working with. Is duplicate content a concern in this case? What measures are recommended to help preserve their North American search inclusion while not serving as a detriment to external (European/Asian CCTLDs) properties aimed for other geos/languages? EDIT: After posting, this was read. Any thoughts? http://searchengineland.com/google-webmaster-tools-provides-details-on-duplicate-content-across-domains-99246
International SEO | | eMagineSEO0