Avoiding "Duplicate Page Title" and "Duplicate Page Content" - Best Practices?
-
We have a website with a searchable database of recipes. You can search the database using an online form with dropdown options for:
Course (starter, main, salad, etc)
Cooking Method (fry, bake, boil, steam, etc)
Preparation Time (Under 30 min, 30min to 1 hour, Over 1 hour)Here are some examples of how URLs may look when searching for a recipe:
find-a-recipe.php?course=starter
find-a-recipe.php?course=main&preperation-time=30min+to+1+hour
find-a-recipe.php?cooking-method=fry&preperation-time=over+1+hourThere is also pagination of search results, so the URL could also have the variable "start", e.g.
find-a-recipe.php?course=salad&start=30
There can be any combination of these variables, meaning there are hundreds of possible search results URL variations.
This all works well on the site, however it gives multiple "Duplicate Page Title" and "Duplicate Page Content" errors when crawled by SEOmoz.
I've seached online and found several possible solutions for this, such as:
- Setting canonical tag
- Adding these URL variables to Google Webmasters to tell Google to ignore them
- Change the Title tag in the head dynamically based on what URL variables are present
However I am not sure which of these would be best.
As far as I can tell the canonical tag should be used when you have the same page available at two seperate URLs, but this isn't the case here as the search results are always different.
Adding these URL variables to Google webmasters won't fix the problem in other search engines, and will presumably continue to get these errors in our SEOmoz crawl reports.
Changing the title tag each time can lead to very long title tags, and it doesn't address the problem of duplicate page content.
I had hoped there would be a standard solution for problems like this, as I imagine others will have come across this before, but I cannot find the ideal solution.
Any help would be much appreciated.
Kind Regards
-
I agree that you shouldn't try to get search pages indexed.
Google loves clean URL's and "?course=starter" is not really a clean URL.One of the reasons why Google doesn't like this is because it's not good for the user experience.
Nobody is gonna remember a url with ? and = in it.My recommendation would definitely be to change the structure of the website that every recipe has it's own page and that these recipes are sorted into categories. Then you can still use the search function so that people will be able to search and they will find their recipes. Trying to get these pages higher in Google may help in some ways, but i think it will serve you better to focus on getting pages with a clean URL structure higher in Google.This should also improve your CTR (Click Through Rate) because people trust websites more if they understand the URL.
About the duplicate content issues. Search pages are not really meant to be indexed by search engines in my experience. I always no-index the search pages because of the possible duplicate content issues and they are custom pages which appear bases on what a user does. Not something i would want to get found on in the search engines.
I hope I helped, if anything is unclear or you need more advice or have a different question than please let me know.
-
I read all the replies and question is still unanswered.
I have the exact same problem with WordPress when using nextpage tag to split a single post into multiple pages.
Wordpress SEO plugin from Yoast doesn't handle this. If I create a blog one post with 3 more pages, url structure looks like the following
With Wordpress SEO Plugin, I can set Title and Description for just ONCE ( not for all 4 pages). This results in Duplicate Title Tags.
Possible Solution 1: Re-write Title using wp_title if page is >= 2. Appending page number to the Title could fix this.
Possible Solution 2: Add www.example.com/post-link/ as canonical link in Advanced Tab of Wordpress SEO
-
smaavie, are you going pick one of these fine answers to mark this Q&A as "Answered"?
-
For few of the page title has warning for my blog campgain has duplicate title !
We used Photoshop tutorials, when we create new article on different new techniques ! what i can't change them ! Because traffic and social votes will be affected.
WordPress slug Changing will become 404 error to the medium and high PA . Before using All in seo Plugin, I don't know that we should use short length url ! so i used the same title in the url ? i don't want to change it ?
any solution ?
-
As this question is rather old but still is marked as "unanswered" I take the liberty to post an answer this late which I hope not only you will benefit from but all other webmasters/SEOs with similar issues.
First of all: Duplicate pages and therefrom duplicate title-tags etc. are of course meant to be taken serious but there are no easy fixes in my opinion and especially not if your design and database set-up is causing large amounts of duplicated content.
But is it a big problem Google-wise? I have my doubts based on conflicting signs and indications given by Google Webmaster Tools and the SERPs in general.
An example I just dug up for you: One of the big players in the field of recipes (allrecipes.com) has +5.000 search results indexed by Google ALL with duplicate title-tags which would bring up all the red flags in the Dashboard.
But based on Googles trends for searches, allrecipes.com are still outperforming their closest competitors. Their search result-urls are unique but all have the same title-tag...so +5.000 duplicate title-tags is probably not really a problem in their perspective.
What to do then?
All though your website seems to have been designed with quite a few potential problems built into its core I would personally hesitate spending a lot of resources on fixing it, especially if your traffic from Google is not taking a bashing.
Eventually your website will be in need for a design re-do and perhaps a change of content management system/database system.
Plan ahead and make sure that you will be able to control this issue in your next version.
It could f. i. be by having all search results appear more like individual pages with individual urls. With a little bit of effort you could make each search result unique with unique title tag and url and thereby bring more traffic to your site.
Best of luck with your efforts
Jens Peter
-
Every site you monitor should have a keyword distribution sheet in excel.
Each line will have the url, meta description, title tag, and h1 tag showing (LEN) character count for meta details. With a further column showing the keywords targeted for each page.
With this you have a way to monitor each page in a more direct visual way and avoid duplication, especially titles and meta descriptions.
-
The best way to fix the issue is address it from the server level - so page level creation and urls.
Link canonical is good, but is really a tier three level fix.
Starting at the root is best. You will want to ensure you have:
1. A logical taxonomy, which is a breakdown of the core topic into sub-categories for classification purposes
2. A logical way to tag categories and entities with meaningful tags, or search based on title, content (tags or keys work well - a programmer should be able to help with this)
3. Rewriting the urls as was mentioned so that any urls exported are always exact urls and not using variables or queries
4. 301 redirect appropriate core query urls to the new urls, and implement internal links to the new urls to reinforce that content and show search engines that it is priority.
5. Continue to run reports regularly and monitor the amount of duplicate content. -
Solution:
-
Download "All in one SEO" plugin
-
Go to plugin settings. Check these settings:
Use no inxex for -categories
-tags
-search (!)
-Achieves
That will prevent duplicate content issues if you use Wordpress.
-
-
Setting canonical tags would be the way I would go, but make sure you have got good seo on the rest of the site for the recipies etc.
-
Take Lonnie's advice. Install Yoast. Use the rel=next tags that the software inserts for you automatically. Yoast will fix it all.
Another WP plugin is called htaccess control. And it is also used for this same purpose... It's a little simpler than the Yoast plugin, and if you already have an SEO plugin you like-- or worst yet-- it is built into your theme, like Thesis...
Just go with htaccess control. It is simple to use and your problems will be solved in minutes.
-
- Setting canonical tag, you should be already doing this rather it's a problem here or not as outlined in seomoz tools you may be missing out on link juice.
I agree with this guy, but I would like to add, why do you want google to crawl your searchable index? Aren't all of your recipes found on your site already by picking categories from a menu?
-
The best way to handle this is via the URL Parameters Setting in Google Webmaster or a robots.txt file.
Google added this functionality to handle the exact issues your'e describing, so there's no need to drastically change functionalities which would likely require editing core files in your CMS.
If you click on URL Parameters under Site Configuration in Google Webmaster you will find a list of queries and for each one there available options that instruct google as to how to handle these pages.
To do this:
1. Click Edit for the Paramater you'd like to configure (i.e. course, cooking, etc).
2. In the Dropdown Menu, select Yes. Changes, reorders, or narrow page contents.
3. Choose the option that best describes how the parameter affects the page content.
4. Choose how GoogleBot should crawl these pages.
- I usually choose "Let GoogleBot Decide" as it's Google your trying to please ;). I've designed and optimized several eCommerce store with multiple parameters and this option handles the crawling and indexing of these pages correctly 99% of the time. If you still experience Duplicate Content issues after editing these settings, simply choose the Ignore option.
Dynamic websites are very common these days and this tool is designed by Google specifically to handle parameters in the best possible way and allow Google to understand the URL structure of your site.. The "Don't have dynamic URLs solution" isn't a solution at all, as many modern functionalities rely on dynamic URLs, such as layered navigation in Magento or other eCommerce platforms. How do you suggest filtering products by price, size, color, etc without creating dynamic URLs? These functionalities IMPROVE user experience and navigation. The text in the address bar isn't always the important factor when a user is navigating a site.
Don't overthink it.
Take advantage of the functionality and only de-index pages that are causing duplicate content problems. If you notice specific dynamic URLs are appearing in SERPs too often then create a 301 redirect from that dynamic URL to a landing page with more user friendly URL.
Hope this helps.
Anthony
-
I agree it is best to get the individual pages indexed. Dont have dynamic pages. Instead come up with categories that make sense, have them indexed.
-
This is something that I've been working on lately. I've been really successfully avoiding duplicate content by using canonical linking, however this has not solve the duplicate Titles nor the duplicate meta descriptions. If you are using a normal web site (static) to post your content as single pages manually, then your only concern would fall into the search pages.
I've switched 100% to Wordpress Blog platforms because of two reasons.
- Google loves them better
- Easier to control content
I've been very successful avoiding duplicate content except for three areas but I do have the solution to repair them as well and I'm currently taking on this task.
The 3 areas of concern are:
- Duplicate Titles
- Duplicate meta descriptions
- Scrapers snatching my unique content and making them their own.
The 3 solutions are: (wordpress platform)
- Duplicate Titles are furnished by pagination next/previous or page #'s at the bottom of each page.
Although wordpress hasn't included this function within the core of its platform yet, Wordpress SEO by Yoast (plugin) automatically add's the new suggested syntax by Google.
Enter rel="next" and rel="prev"
Now, as it goes with these things, Google has just posted the solution. They've asked to add
rel="next"
andrel="prev"
to paginated archives, so that they can distinguish them as a series and, quote: Send users to the most relevant page/URL—typically the first page of the series.The above syntax will solve our pagination duplicate titles and search paginations. The plugin also adds tag terms at the end of the title for each page. This makes the Title unique.
-
Now the above also tells Google that page #1 is the canonical Title and meta description for all paginations, therefore your meta description is now accurate and safe. The plugin also has an advanced feature which allows you to provide a different description per page other than what the page actually states. Making this slight change makes all the difference.
-
The next problem is robbery or copy infringing my content. My unique content has been scraped and posted without my permission, however now... we can use another rel= syntax to point the article back to the original owner.
rel=”author” and rel=”me” in WP and other platforms
You can allow people to use your content, however the rel="me" tells search engines who the unique content really belongs to. and the rel="author" points to me as well.
This attribute allows you to tell Google who you are as an author and what articles you write. Google has indicated that they believe the authority of an author may even be weighted more heavily than traditional on page metrics, like page or domain authority. As Matt Cutts stated at SMX West, “The concept is that if an author is trustworthy, why does it matter what site the article appears on?”. Author authority also has implications for the impending Panda 2.2 update, which will affect the sites that steal content from other sites to post on their own. If Google sees the same article on 10 different sites, and 1 of those sites clearly identifies an author, marked up with the "rel=author" attribute, which site do you think Google is going to rank?
This is the extent of my research on the above and so far its working well. I hope the above helps for you too.
Cheers!
-
Personally I believe its best practice to have user friendly urls, rather than search generated ones. Google favours this and so do the users. It may be a lot more work to implement, but in my experience (having a site with a lot of categories and posts) it was well worth it.
-
Thanks Keri,
Our current experience is that search results from our site are showing up in Google results, sometimes quite high.
So, I'm reluctant to change anything too drastically - "if it ain't broke, don't fix it". But ... maybe we could get slightly higher rankings if we made some minor alterations?
Is there any 'best practice' guidance I could look at to learn more about this specific issue?
Thanks for your help.
David
-
I think Baptiste is referring to Google's preference for not including search results in their search results, as the URL in the example appeared to be a search result.
-
Hello Baptiste,
I'm keen to know more about why you believe we would get penalised for this. What, specifically, should we seek to avoid in order to avoid the penalty?
Thanks for your help
David
-
I noticed this question is still listed as unanswered. Did you come up with a solution you can share with us, and any information about how well it worked? Or are you still looking for advice? Would be great if you could pop back in with an update. Thanks!
-
I manage websites specialising in holiday rentals, so the search pages are very powerful however I only use these for customer experience. For my seo I create pages based on the areas, types of properties, specific searches i.e. villas in florida etc..
I think when building websites you must always have two outlooks; users & seo
-
Why do you think this? Is it Part of googles terms of service?
-
Hope you don't create link from visitor's researches, like find-a-recipe.php?course=salad**&q=tomatoes** as you would get penalised !
-
Sure, I understand where you're coming from. I still think there's no easy solution to this, but maybe someone else will have some interesting suggestions.
What I was suggesting in my first reply above is pretty much in line with what Baptiste is saying below. Google used to be very tough on people trying to index search results pages and that's why personally I would try going a bit different way.
Cheers!
-
I would suggest making indexable pages for courses, the rest of the parameters are rather user orientated and - I think - not usefull for SEO. This means separating the search script with browse pages.
This means making find-a-recipe.php, which looks like the search engine, forbidden to robots. Instead, you should have a category browser, using only the course (I suppose no recipe have multiple courses ?). You would have url like :
/recipes/ => all recipes, paginated
/recipes/start/ => all starter recipes, paginated
/recipes/starter/fry/ => fried starter recipes, but you should check the search volume of those expression, like "fried starter recipes". If you have a very small volume of recipes, wait until every subpages of /recipes/starter/ have at least 5 recipes.
The goal here is to make your recipe pages easy to index, with a strong focus on the course type. Although the course may not be the best root category for recipe, this should be a good way to may your site seo friendly.
-
Thanks Nemek,
I appreciate your answer.
However, as the site owner my instinct is to seek to get as many pages as possible indexed, so I'd like to get further advice about this before I take action.
The search results pages on our site often mirror what people are specifically searching for in Google, so we'd love our results pages to be highly ranked so as to help these people find what they want, quickly.
Does anyone else have an opinion on the best way forward for us?
Thanks in advance.
-
Technically, the search engines don't want to crawl other "search results". Personally I would try to get individual pages and category pages indexed, while avoid trying to index and "canonicalize" search result pages.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Consolidating a Large Site with Duplicate Content
I will be restructuring a large website for an OEM. They provide products & services for multiple industries, and the product/service offering is identical across all industries. I was looking at the site structure and ran a crawl test, and learned they have a LOT of duplicate content out there because of the way they set up their website. They have a page in the navigation for “solution”, aka what industry you are in. Once that is selected, you are taken to a landing page, and from there, given many options to explore products, read blogs, learn about the business, and contact them. The main navigation is removed. The URL structure is set up with folders, so no matter what you select after you go to your industry, the URL will be “domain.com/industry/next-page”. The product offerings, blogs available, and contact us pages do not vary by industry, so the content that can be found on “domain.com/industry-1/product-1” is identical to the content found on “domain.com/industry-2/product-1” and so-on and so-forth. This is a large site with a fair amount of traffic because it’s a pretty substantial OEM. Most of their content, however, is competing with itself because most of the pages on their website have duplicate content. I won’t begin my work until I can dive in to their GA and have more in-depth conversations with them about what kind of activity they’re tracking and why they set up the website this way. However, I don’t know how strategic they were in this set up and I don’t think they were aware that they had duplicate content. My first thought would be to work towards consolidating the way their site is set up, so we don’t spread the link-equity of “product-1” content, and direct all industries to one page, and track conversion paths a different way. However, I’ve never dealt with a site structure of this magnitude and don’t want to risk messing up their domain authority, missing redirect or URL mapping opportunities, or ruin the fact that their site is still performing well, even though multiple pages have the same content (most of which have high page authority and search visibility). I was curious if anyone has dealt with this before and if they have any recommendations for tackling something like this?
On-Page Optimization | | cassy_rich0 -
Duplicate page titles and Content in Woocommerce
Hi Guys, I'm new to Moz and really liking it so far!
On-Page Optimization | | jeeyer
I run a eCommerce site on Wordpress + WooCommerce and ofcourse use Yoast for SEO optimalisation I've got a question about my first Crawl report which showed over 600 issues! 😐 I've read that this is something that happens more often (http://moz.com/blog/setup-wordpress-for-seo-success). Most of them are categorized under:
1. Duplicate Page Titles or;
2. Duplicate Page Content. Duplicate Page Titles:
These are almost only: product category pages and product tags. Is this problem beeing solved by giving them the right SEO SERP? I see that a lot of categories don't have a proper SEO SERP set up in yoast! Do I need to add this to clear this issue, or do I need to change the actual Title? And how about the Product tags? Another point (bit more off-topic) I've read here: http://moz.com/community/q/yoast-seo-plugin-to-index-or-not-to-index-categories that it's advised to noindex/follow Categories and Tags but isn't that a wierd idea to do for a eCommerce site?! Duplicate Page Content:
Same goes here almost only Product Categories and product tags that are displayed as duplicate Page content! When I check the results I can click on a blue button for example "+ 17 duplicates" and that shows me (in this case 17 URLS) but they are not related to the fist in any way so not sure where to start here? Thanks for taking the time to help out!
Joost0 -
Duplicate Home Page
Hi, I have a question around best practise on duplicate home pages. The /index.aspx page is showing up as a top referrer in my analytics. I have the rel=canonical tag implemented for the www.mysite.com on both pages. Do I need to 301 the /index.aspx to the mysite.com? I have a lot of links pointing to the /index.aspx (half of those are coming from the mysite.com). www.mysite.com/index.aspx www.mysite.com Many thanks Jon
On-Page Optimization | | JonRaubenheimer0 -
"irrelevant pages of a site"
Hi there! Some pages of my site like "contact" or "registration": Should they have a title and a description tag? They are pages that I don't want them to be shown in the SERPs....Could I be penalized by google If I don't do so? The SEOMOZ crawling tool warms me about this issue (to short titles, no meta-description tags....) Many thanks
On-Page Optimization | | juanmiguelcr0 -
How to fix joomla duplicate page titles??
Hello, I have been using seomoz and sh404sef for almost a month now and can't seem to figure out my duplicate page title issue on my joomla site. I have changed everything I thought necessary but no luck. I have one page that I am trying to make dissapear from duplicate page title error and if I can figure it out i'm sure I can use same fix for the other 1,000 duplicate page titles being reported in my seomoz crawl report I have one link that I am working with sef url: http://www.mysite.com/contactus.html I have used sh404sef url manager and added page title, description and keywords to this link and for 3 weeks straight now it still shows that it is duplicate page title in my seomoz crawl report Is there something that I am missing somewhere or did i setup a bad campaign that is looking at wrong things on site? I 've attached image of my sh404sef settings for this link Thanks in advance screenshot1.jpg
On-Page Optimization | | 41global0 -
How should I rephrase these pages to avoid Phrase duplication within Title Tags
How should I rephrase these pages to avoid Phrase duplication within Title Tags Duplicate Page Title Page1-http://organicfruitbasketsflorist.com/index-2.html Page2- http://organicfruitbasketsflorist.com/Fruit_Baskets_Organic_Fruit_Baskets_New_York_NY.html Page3- http://organicfruitbasketsflorist.com/Fruit_Baskets_Edible_Fruit_Baskets_New_York_NY.html Page4organicfruitbasketsflorist.com/Fruit_Baskets_Business_Fruit_Baskets_New_York_NY.html Page5-http://organicfruitbasketsflorist.com/Fruit_Baskets_Fresh_Flowers_Delivered_New_York_NY.html Page6-http://organicfruitbasketsflorist.com/Coupons.htmlAmi
On-Page Optimization | | amydiamond0 -
Replacing "_" with "-" in url, results in new url?
We ran SEOmoz's "On-Page Optimization" tool on a url which contains the character "_". According to the tool: "Characters which are less commonly used in URLs may cause problems with accessibility, interpretation and ranking in search engines. It is considered a best practice to stick to standard URL structures to avoid potential problems." "Rewrite the URL to contain only standard characters." Therefore we will rewrite the url, replacing "_" with "-". Will search engines consider the "-" url a different one? Do we need to 301 the old url to the new one? Thanks for your help!
On-Page Optimization | | gerardoH0 -
Duplicate page title issues with a CMS
I am using MODx as a CMS on a site and trying to eliminate duplicate page titles. url.com/ url.com/[~897~] which is really {~897~} its a resource number. url.com/home/ How can I resolve this issue when its all one page in the CMS? thanks
On-Page Optimization | | tjsherrill0