Is this duplicate content?
-
My client has several articles and pages that have 2 different URLs
For example:
is the same article as:
I was not sure if this was duplicate content or not ...
Or if I should be putting "/article.cfm" into the robots.txt file or not..
if anyone could help me out, that would be awesome!
Thanks
-
Agreed - although I think a 301-redirect or canonical tag implementation would probabyl be ok. If there's a database lookup that can translate the DocID into a URL string, the canonical is easy (I write some CF code, so I can at least tell you it's doable). Keep in mind that "article.cfm" is only one template, so if you can find a solution that's data-driven, it's just as easy for 1,000 pages as it is for 10.
You could also create a dynamic 301-redirect via <cfheader>- the core logic is the same. Basically, you look up the URL from the DocID and dynamically create the tag. You just need someone who understands your CMS and data. The actual code is only a few lines, but understanding your setup is the time-consuming part.</cfheader>
-
ATMOS, those are just the same page, so Canonical tag should do it, but also you want to stop google indexing it, so you could detect that it is called with the article.cfm and use a no index META tag too, but not if it uses the friendly url
-
I mostly agree with kchan.
- It is considered duplicate content.
- Simplest way is to do rel canonical for the pages with ids.
However, I suspect 301 redirect is not the best way. Especially, if your website is using Omniture and/or Google web analytics code, you might get miscalculated traffic through them because of 301 redirect.
Be careful if you choose the last route.
-
Awesome Chan, thanks. That was my thought as well. Most difficult part will be determining how to get that script in place.
-
Any chance you can spend a little time writing it out?
My guess is that we should be doing a rel canonical tag on all the article.cfm?intDocID=22572 type pages, that would then direct the bots to our /bc-blazes-construction-trail. But what's the easiest way to do that across the whole site?
-
Hello,
It sure is duplicate content. By putting "/article.cfm" into the robot.txt won't work because if you do that you are just re-directing the whole folder. You need to do a permanent re-direct. I had a brief look at the site and it seems like there are over 1000+ pages. This might take a while but it is neccessary to do it, if not your clients rankings will not perform and most likely penalised. A simple way would be doing a canonical link in /article.cfm?intDocID=22572 so you are showing google the main article is located at /bc-blazes-construction-trail.
However, the best way would be doing a 301 permanent re-direct of course. I'm sure you could get a web dev to write a script to automatically run through the database and output the re-directs than manually re-directing 1000+ pages. If it can't be done, you could outsource it on freelancer.com for around $2-300.
Thanks
-
That would definitely be considered duplicate content. There are a few things you can do to fix it, but rather than wasting a bunch of time writing it out here I would recommend visiting the link below for more detailed info:
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Penalty for duplicate content on the same website?
Is it possible to get a penalty for duplicate content on the same website? I have a old custom-built site with a large number of filters that are pre-generated for speed. Basically the only difference is the meta title and H1 tag, with a few text differences here and there. Obviously I could no-follow all the filter links but it would take an enormous amount of work. The site is performing well in the search. I'm trying to decide whether if there is a risk of a penalty, if not I'm loath to do anything in case it causes other issues.
Intermediate & Advanced SEO | | seoman100 -
Duplicate content across different domains
Hi Guys, Looking for some advice regarding duplicate content across different domains. I have reviewed some previous Q&A on this topic e.g. https://moz.com/community/q/two-different-domains-exact-same-content but just want to confirm if I'm missing anything. Basically, we have a client which has 1 site (call this site A) which has solids rankings. They have decided to build a new site (site B), which contains 50% duplicate pages and content from site A. Our recommendation to them was to make the content on site B as unique as possible but they want to launch asap, so not enough time. They will eventually transfer over to unique content on the website but in the short-term, it will be duplicate content. John Mueller from Google has said several times that there is no duplicate content penalty. So assuming this is correct site A should be fine, no ranking losses. Any disagree with this? Assuming we don't want to leave this to chance or assume John Mueller is correct would the next best thing to do is setup rel canonical tags between site A and site B on the pages with duplicate content? Then once we have unique content ready, execute that content on the site and remove the canonical tags. Any suggestions or advice would be very much appreciated! Cheers, Chris
Intermediate & Advanced SEO | | jayoliverwright0 -
Duplicate content but different pages?
Hi there! Im getting LOTS of "duplicate content" pages but the thing is they are different pages. My website essentially is a niche video hosting site with embedded videos from Youtube. Im working on adding personal descriptions to each video but keeping the same video title (should I re-word it from the original also? Any help?
Intermediate & Advanced SEO | | sarevme0 -
Duplicate content on URL trailing slash
Hello, Some time ago, we accidentally made changes to our site which modified the way urls in links are generated. At once, trailing slashes were added to many urls (only in links). Links that used to send to
Intermediate & Advanced SEO | | yacpro13
example.com/webpage.html Were now linking to
example.com/webpage.html/ Urls in the xml sitemap remained unchanged (no trailing slash). We started noticing duplicate content (because our site renders the same page with or without the trailing shash). We corrected the problematic php url function so that now, all links on the site link to a url without trailing slash. However, Google had time to index these pages. Is implementing 301 redirects required in this case?1 -
Duplicate content issue with pages that have navigation
We have a large consumer website with several sections that have navigation of several pages. How would I prevent the pages from getting duplicate content errors and how best would I handle SEO for these? For example we have about 500 events with 20 events showing on each page. What is the best way to prevent all the subsequent navigation pages from getting a duplicate content and duplicate title error?
Intermediate & Advanced SEO | | roundbrix0 -
Is a different location in page title, h1 title, and meta description enough to avoid Duplicate Content concern?
I have a dynamic website which will have location-based internal pages that will have a <title>and <h1> title, and meta description tag that will include the subregion of a city. Each page also will have an 'info' section describing the generic product/service offered which will also include the name of the subregion. The 'specific product/service content will be dynamic but in some cases will be almost identical--ie subregion A may sometimes have the same specific content result as subregion B. Will the difference of just the location put in each of the above tags be enough for me to avoid a Duplicate Content concern?</p></title>
Intermediate & Advanced SEO | | couponguy0 -
Penalized for Similar, But Not Duplicate, Content?
I have multiple product landing pages that feature very similar, but not duplicate, content and am wondering if this would affect my rankings in a negative way. The main reason for the similar content is three-fold: Continuity of site structure across different products Similar, or the same, product add-ons or support options (resulting in exactly the same additional tabs of content) The product itself is very similar with 3-4 key differences. Three examples of these similar pages are here - although I do have different meta-data and keyword optimization through the pages. http://www.1099pro.com/prod1099pro.asp http://www.1099pro.com/prod1099proEnt.asp http://www.1099pro.com/prodW2pro.asp
Intermediate & Advanced SEO | | Stew2220 -
Could you use a robots.txt file to disalow a duplicate content page from being crawled?
A website has duplicate content pages to make it easier for users to find the information from a couple spots in the site navigation. Site owner would like to keep it this way without hurting SEO. I've thought of using the robots.txt file to disallow search engines from crawling one of the pages. Would you think this is a workable/acceptable solution?
Intermediate & Advanced SEO | | gregelwell0