404 for duplicate content?
-
Sorry, I think this is my third question today...
But I have a lot of duplicated content on my site. I use joomla so theres a lot of unintentional duplication. For example, www.mysite.com/index.php exists, etc.
Up till now, I thought I had to 301 redirect or rel=canonical these "duplicated pages."
However, can I just 404 it? Is there anything wrong with this rpactice in regards to SEO?
-
I agree with Andy here. Too many 404's can hurt your site. EVEN Google says that in GWT. I wouldn't do any 404s. I would 301 or robot.txt folders.
You may want to robots.txt some folders. Sometimes you can get a plugin and fix things quickly.
-
Hi Kyu,
Remember, canonical is only a suggestion to google of which page should be delivering the content - it is still up to them what they do. In practice though, this is what many opt for.
301's are a permanent redirect and too many can suggest a poor underlying site - you wouldn't want a 301 for every page if there were a lot of them.
You could also think about Robots to remove some of the duplicated pages so they never get spidered, or just no-index them.
404's for me wouldn't be the ideal scenario because somewhere in the site, it can lead to what is basically a dead page. Too many 404's can actually harm your ranking because when Google spiders and finds them, if you have a large enough site, they could be met with 200+ dead pages!
On some sites, you are able to just remove the pages altogether, but you can't do this with the likes of Joomla.
Think about no-indexing / robots because although the pages will still be there, you are telling Google not to bother. This is the route many SEO's are taking now.
Andy
-
You are very welcome. I think "simpler" could be a relative term All three are appropriate in different situations. However, there are times when people have very limited access to source code or to the backends of their websites, so then one solution might work better than another.
As far as 404s go it's really all about what's best and most appropriate from a user standpoint. If you can guide visitors to content relevant to their search query via a 301-redirect, they are probably going to be more satisfied with that than a 404. This could potentially indirectly effect your SEO because if your bounce rate increases or your 404 pages results in a lot of pogo-sticking by potential visitors, your site could be effected negatively by Googe's algorithm.
When at all possible, I try to do a 301-redirect. But in the cases of really old content that may no longer accurately represent our content or products (and that also doesn't have veyr many inbound links) a 404 might be just fine.
Sorry, that's a bit of a long answer, but I hope it helps!
Dana
-
Thanks Dana! Youve been so helpful!
But one thing I am confused about, when i read articles about how to fix duplicate content, they always talk about the best two options being 301 or rel=canonical. Why is that?
Isnt 404 error simpler?
Hmm, or is 404 just simpler in my case beacuse all my duplicated pages are pages that users will never go to?
-
Yes, you could allow those pages to 404 and in some instances that may be preferable to you. No, there is no negative effect on SEO from 404's. The only negative impact is really on your users. To minimize this, you might consider creating a nice, friendly, customer 404 page instead of using Google's defult 404 error page. Hope that helps!
Dana
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Shall we add engaging and useful FAQ content in all our pages or rather not because of duplication and reduction of unique content?
We are considering to add at the end of alll our 1500 product pages answers to the 9 most frequently asked questions. These questions and answers will be 90% identical for all our products and personalizing them more is not an option and not so necessary since most questions are related to the process of reserving the product. We are convinced this will increase engagement of users with the page, time on page and it will be genuinely useful for the visitor as most visitors will not visit the seperate FAQ page. Also it will add more related keywords/topics to the page.
Intermediate & Advanced SEO | | lcourse
On the downside it will reduce the percentage of unique content per page and adds duplication. Any thoughts about wether in terms of google rankings we should go ahead and benefits in form of engagement may outweight downside of duplication of content?0 -
Identifying Duplicate Content
Hi looking for tools (beside Copyscape or Grammarly) which can scan a list of URLs (e.g. 100 pages) and find duplicate content quite quickly. Specifically, small batches of duplicate content, see attached image as an example. Does anyone have any suggestions? Cheers. 5v591k.jpg
Intermediate & Advanced SEO | | jayoliverwright0 -
Lot of duplicate content and still traffic is increasing... how does it work?
Hello Mozzers, I've a dilemma with a client's site I am working on that is make me questioning my SEO knowledge, or the way Google treat duplicate content. I'll explain now. The situation is the following: organic traffic is constantly increasing since last September, in every section of the site (home page, categories and product pages) even though: they have tons of duplicate content from same content in old and new URLs (which are in two different languages, even if the actual content on the page is in the same language in both of the URL versions) indexation is completely left to Google decision (no robots file, no sitemap, no meta robots in code, no use of canonical, no redirect applied to any of the old URLs, etc) a lot (really, a lot) of URLs with query parameters (which brings to more duplicated content) linked from the inner page of the site (and indexed in some case) they have Analytics but don't use Webmaster Tools Now... they expect me to help them increase even more the traffic they're getting, and I'll go first on "regular" onpage optimization, as their title, meta description and headers are not optimized at all according to the page content, but after that I was thinking on fixing the issues with indexation and content duplication, but I am worried I can "break the toy", as things are going well for them. Should I be confident that fixing these issues will bring to even better results or do you think is better for me to focus on other kind of improvements? Thanks for your help!
Intermediate & Advanced SEO | | Guybrush_Threepw00d0 -
How to resolve duplicate content issues when using Geo-targeted Subfolders to seperate US and CAN
A client of mine is about to launch into the USA market (currently only operating in Canada) and they are trying to find the best way to geo-target. We recommended they go with the geo-targeted subfolder approach (___.com and ___.com/ca). I'm looking for any ways to assist in not getting these pages flagged for duplicate content. Your help is greatly appreciated. Thanks!
Intermediate & Advanced SEO | | jyoung2220 -
Duplicate Content Question
We are getting ready to release an integration with another product for our app. We would like to add a landing page specifically for this integration. We would also like it to be very similar to our current home page. However, if we do this and use a lot of the same content, will this hurt our SEO due to duplicate content?
Intermediate & Advanced SEO | | NathanGilmore0 -
How are they avoiding duplicate content?
One of the largest stores in USA for soccer runs a number of whitelabel sites for major partners such as Fox and ESPN. However, the effect of this is that they are creating duplicate content for their products (and even the overall site structure is very similar). Take a look at: http://www.worldsoccershop.com/23147.html http://www.foxsoccershop.com/23147.html http://www.soccernetstore.com/23147.html You can see that practically everything is the same including: product URL product title product description My question is, why is Google not classing this as duplicate content? Have they coded for it in a certain way or is there something I'm missing which is helping them achieve rankings for all sites?
Intermediate & Advanced SEO | | ukss19840 -
How to resolve Duplicate Page Content issue for root domain & index.html?
SEOMoz returns a Duplicate Page Content error for a website's index page, with both domain.com and domain.com/index.html isted seperately. We had a rewrite in the htacess file, but for some reason this has not had an impact and we have since removed it. What's the best way (in an HTML website) to ensure all index.html links are automatically redirected to the root domain and these aren't seen as two separate pages?
Intermediate & Advanced SEO | | ContentWriterMicky0 -
Should I robots block site directories with primarily duplicate content?
Our site, CareerBliss.com, primarily offers unique content in the form of company reviews and exclusive salary information. As a means of driving revenue, we also have a lot of job listings in ouir /jobs/ directory, as well as educational resources (/career-tools/education/) in our. The bulk of this information are feeds, which exist on other websites (duplicate). Does it make sense to go ahead and robots block these portions of our site? My thinking is in doing so, it will help reallocate our site authority helping the /salary/ and /company-reviews/ pages rank higher, and this is where most of the people are finding our site via search anyways. ie. http://www.careerbliss.com/jobs/cisco-systems-jobs-812156/ http://www.careerbliss.com/jobs/jobs-near-you/?l=irvine%2c+ca&landing=true http://www.careerbliss.com/career-tools/education/education-teaching-category-5/
Intermediate & Advanced SEO | | CareerBliss0