Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How can I avoid duplicate content for a new landing page which is the same as an old one?
-
Hello mozers!
I have a question about duplicate content for you...
One on my clients pages have been dropping in search volume for a while now, and I've discovered it's because the search term isn't as popular as it used to be. So... we need to create a new landing page using a more popular search term.
The page which is losing traffic is based on the search query "Can I put a solid roof on my conservatory" this only gets 0-10 searches per month according to the keyword explorer tool. However, if we changed this to "replacing conservatory roof with solid roof" this gets up to 500 searches per month. Muuuuch better!
The issue is, I don't want to close down and re-direct the old page because it's got a featured snippet and sits in position 1. So I'd like to create another page instead... however, as the two are effectively the same content, I would then land myself in a duplicate content issue.
If I were to put a rel="canonical" tag in the original "can I put a solid roof...." page but say the master page is now the new one, would that get around the issue?
-
@Virginia-Girtz To avoid duplicate content issues when creating a new landing page that is similar to an old one, consider the following strategies:
-
301 Redirect: If the old landing page is no longer needed, you can redirect its URL to the new landing page using a 301 redirect. This tells search engines that the old page has permanently moved to the new location.
-
Canonical Tags: Implement canonical tags on the new landing page pointing to the old landing page URL. This informs search engines that the content on the new page is a duplicate of the old page and should be indexed under the old page's URL.
-
Content Variation: Rewrite the content on the new landing page to make it sufficiently different from the old one. This could involve changing the wording, adding new information, or altering the layout.
-
Noindex Tag: If the old landing page is still relevant but you want to prioritize the new one, you can use a noindex tag on the old page. This prevents search engines from indexing the old page while still allowing users to access it.
-
Consolidate Content: Consider consolidating the content from both landing pages into a single, comprehensive page. This helps avoid duplication and can improve user experience by providing all relevant information in one place.
-
Robots.txt: Use the robots.txt file to block search engines from crawling one of the landing pages. However, this approach should be used cautiously as it may also prevent search engines from discovering other valuable content on your site.
I apply all these experiment on this of my client site
By implementing one or a combination of these strategies, you can effectively address duplicate content concerns while maintaining the visibility and relevance of your landing pages.
-
-
So what you want for every page and blog post on your website is unique, high-quality white hat content marketing.
We applied this white hat SEO method to a U.K garden room company, website and after we rewrote the pages, the organic visitor numbers increased.
-
What I've usually seen with canonicals is that Google either removes the noncanonical page from its index, or it ignores your canonical and treats them as two separate pages. I haven't seen an example where a canonical lets you get the best of both worlds.
I agree with Nozzle - you can tweak your existing content to target both phrases! Google understands synonyms, so if anything, you're just creating a more all around relevant page.
Good luck!
Kristina
-
Since it is effectively the same content you should be able to rank the same page for both phrases.
You just need to include the new keyword within the existing content and test out a few title tag variations to find one that helps you move up the rankings for the new keyword without dropping your ranking for the old keyword.
The first thing I'd test would be to change your title tag from "Can I put a solid roof on my conservatory?" to "Replacing Conservatory Roof with Solid Roof - Can I put a solid roof on my conservatory?". Wait until Google re-crawls the page and check how your rankings fared. If you lose your snippet or drop in rankings for the low volume phrase you can always test out the reverse, "Can I put a solid roof on my conservatory? Replacing Conservatory Roof with Solid Roof", and see what happens then.
Don't be scared to test many variations, even long title tags that seemingly don't follow best practice. You can always change it back to the original and your rankings will go back to what they were before you tested (assuming your competitors didn't gain some awesome back links to overtake you).
Don't mess with the section of content that is being pulled into the featured snippet though so as not to lose that snippet.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Directory with Duplicate content? what to do?
Moz keeps finding loads of pages with duplicate content on my website. The problem is its a directory page to different locations. E.g if we were a clothes shop we would be listing our locations: www.sitename.com/locations/london www.sitename.com/locations/rome www.sitename.com/locations/germany The content on these pages is all the same, except for an embedded google map that shows the location of the place. The problem is that google thinks all these pages are duplicated content. Should i set a canonical link on every single page saying that www.sitename.com/locations/london is the main page? I don't know if i can use canonical links because the page content isn't identical because of the embedded map. Help would be appreciated. Thanks.
Intermediate & Advanced SEO | | nchlondon0 -
New Site (redesign) Launched Without 301 Redirects to New Pages - Too Late to Add Redirects?
We recently launched a redesign/redevelopment of a site but failed to put 301 redirects in place for the old URL's. It's been about 2 months. Is it too late to even bother worrying about it at this point? The site has seen a notable decrease in site traffic/visits, perhaps due to this issue. I assume that once the search engines get an error on a URL, it will remove it from displaying in search results after a period of time. I'm just not sure if they will try to re-crawl those old URLs at some point and if so, it may be worth it to have those 301 redirects in place. Thank you.
Intermediate & Advanced SEO | | BrandBuilder0 -
Is a different location in page title, h1 title, and meta description enough to avoid Duplicate Content concern?
I have a dynamic website which will have location-based internal pages that will have a <title>and <h1> title, and meta description tag that will include the subregion of a city. Each page also will have an 'info' section describing the generic product/service offered which will also include the name of the subregion. The 'specific product/service content will be dynamic but in some cases will be almost identical--ie subregion A may sometimes have the same specific content result as subregion B. Will the difference of just the location put in each of the above tags be enough for me to avoid a Duplicate Content concern?</p></title>
Intermediate & Advanced SEO | | couponguy0 -
Can too many "noindex" pages compared to "index" pages be a problem?
Hello, I have a question for you: our website virtualsheetmusic.com includes thousands of product pages, and due to Panda penalties in the past, we have no-indexed most of the product pages hoping in a sort of recovery (not yet seen though!). So, currently we have about 4,000 "index" page compared to about 80,000 "noindex" pages. Now, we plan to add additional 100,000 new product pages from a new publisher to offer our customers more music choice, and these new pages will still be marked as "noindex, follow". At the end of the integration process, we will end up having something like 180,000 "noindex, follow" pages compared to about 4,000 "index, follow" pages. Here is my question: can this huge discrepancy between 180,000 "noindex" pages and 4,000 "index" pages be a problem? Can this kind of scenario have or cause any negative effect on our current natural SEs profile? or is this something that doesn't actually matter? Any thoughts on this issue are very welcome. Thank you! Fabrizio
Intermediate & Advanced SEO | | fablau0 -
How do I geo-target continents & avoid duplicate content?
Hi everyone, We have a website which will have content tailored for a few locations: USA: www.site.com
Intermediate & Advanced SEO | | AxialDev
Europe EN: www.site.com/eu
Canada FR: www.site.com/fr-ca Link hreflang and the GWT option are designed for countries. I expect a fair amount of duplicate content; the only differences will be in product selection and prices. What are my options to tell Google that it should serve www.site.com/eu in Europe instead of www.site.com? We are not targeting a particular country on that continent. Thanks!0 -
Could you use a robots.txt file to disalow a duplicate content page from being crawled?
A website has duplicate content pages to make it easier for users to find the information from a couple spots in the site navigation. Site owner would like to keep it this way without hurting SEO. I've thought of using the robots.txt file to disallow search engines from crawling one of the pages. Would you think this is a workable/acceptable solution?
Intermediate & Advanced SEO | | gregelwell0 -
Blocking Pages Via Robots, Can Images On Those Pages Be Included In Image Search
Hi! I have pages within my forum where visitors can upload photos. When they upload photos they provide a simple statement about the photo but no real information about the image,definitely not enough for the page to be deemed worthy of being indexed. The industry however is one that really leans on images and having the images in Google Image search is important to us. The url structure is like such: domain.com/community/photos/~username~/picture111111.aspx I wish to block the whole folder from Googlebot to prevent these low quality pages from being added to Google's main SERP results. This would be something like this: User-agent: googlebot Disallow: /community/photos/ Can I disallow Googlebot specifically rather than just using User-agent: * which would then allow googlebot-image to pick up the photos? I plan on configuring a way to add meaningful alt attributes and image names to assist in visibility, but the actual act of blocking the pages and getting the images picked up... Is this possible? Thanks! Leona
Intermediate & Advanced SEO | | HD_Leona0 -
Create new subdomain or new site for new Niche Product?
We have an existing large site with strong, relevant traffic, including excellent SEO traffic. The company wants to launch a new business offering, specifically targeted at the "small business" segment. Because the "small business" customer is substantially different from the traditional "large corporation" customer, the company has decided to create a completely independent microsite for the "small business" market. Purely from a Marketing and Communications standpoint, this makes sense. From an SEO perspective, we have 2 options: Create the new "small business" microsite on a subdomain of the existing site, and benefit from the strong domain authority and trust of the existing site. Build the microsite on a separate domain with exact primary keyword match in the domain name. My sense is that option #1 is by far the better option in the short and long run. Am I correct? Thanks in advance!
Intermediate & Advanced SEO | | axelk0