Google keeps marking different pages as duplicates
-
My website has many pages like this:
mywebsite/company1/valuation
mywebsite/company2/valuation
mywebsite/company3/valuation
mywebsite/company4/valuation
...
These pages describe the valuation of each company.
These pages were never identical but initially, I included a few generic paragraphs like what is valuation, what is a valuation model, etc... in all the pages so some parts of these pages' content were identical.
Google marked many of these pages as duplicated (in Google Search Console) so I modified the content of these pages: I removed those generic paragraphs and added other information that is unique to each company. As a result, these pages are extremely different from each other now and have little similarities.
Although it has been more than 1 month since I made the modification, Google still marks the majority of these pages as duplicates, even though Google has already crawled their new modified version. I wonder whether there is anything else I can do in this situation?
Thanks
-
Google may mark different pages as duplicates if they contain very similar or identical content. This can happen due to issues such as duplicate metadata, URL parameters, or syndicated content. To address this, ensure each page has unique and valuable content, use canonical tags when appropriate, and manage URL parameters in Google Search Console.
-
Yes, there are a few other things you can do if Google is still marking your pages as duplicates after you have modified them to be unique:
-
Check your canonical tags. Canonical tags tell Google which version of a page is the preferred one to index. If you have canonical tags in place and they are pointing to the correct pages, then Google should eventually recognize that the duplicate pages are not actually duplicates.
-
Use the URL parameter tool in Google Search Console. This tool allows you to tell Google which URL parameters it should treat as unique and which ones it should ignore. This can be helpful if you have pages with similar content but different URL parameters, such as pages for different product categories or pages with different sorting options.
-
Request a recrawl of your website. You can do this in Google Search Console. Once Google has recrawled your website, it will be able to see the new, modified versions of your pages.
If you have done all of the above and Google is still marking your pages as duplicates, then you may need to contact Google Support for assistance.
-
-
If Google is marking different pages on your website as duplicates, it can negatively impact your website's search engine rankings. Here are some common reasons why Google may be doing this and steps you can take to address the issue:
Duplicate Content: Google's algorithms are designed to filter out duplicate content from search results. Ensure that your website does not have identical or near-identical content on multiple pages. Each page should offer unique and valuable content to users.
URL Parameters: If your website uses URL parameters for sorting, filtering, or tracking purposes, Google may interpret these variations as duplicate content. Use canonical tags or the URL parameter tool in Google Search Console to specify which version of the URL you want to be indexed.
Pagination: For websites with paginated content (e.g., product listings, blog archives), ensure that you implement rel="next" and rel="prev" tags to indicate the sequence of pages. This helps Google understand that the pages are part of a series and not duplicates.
www vs. non-www: Make sure you have a preferred domain (e.g., www.example.com or example.com) and set up 301 redirects to the preferred version. Google may treat www and non-www versions as separate pages with duplicate content.
HTTP vs. HTTPS: Ensure that your website uses secure HTTPS. Google may view HTTP and HTTPS versions of the same page as duplicates. Implement 301 redirects from HTTP to HTTPS to resolve this.
Mobile and Desktop Versions: If you have separate mobile and desktop versions of your site (e.g., responsive design or m.example.com), use rel="alternate" and rel="canonical" tags to specify the relationship between the two versions.
Thin or Low-Quality Content: Pages with little or low-quality content may be flagged as duplicates. Improve the content on such pages to provide unique value to users.
Canonical Tags: Implement canonical tags correctly to indicate the preferred version of a page when there are multiple versions with similar content.
XML Sitemap: Ensure that your XML sitemap is up-to-date and accurately reflects your website's structure. Submit it to Google Search Console.
Avoid Scraped Content: Ensure that your content is original and not scraped or copied from other websites. Google penalizes sites with duplicate or plagiarized content.
Check for Technical Errors: Use Google Search Console to check for crawl errors or other technical issues that might be causing duplicate content problems.
Structured Data: Ensure that your structured data (schema markup) is correctly implemented on your pages. Incorrectly structured data can confuse search engines.
Regularly monitor Google Search Console for any duplicate content issues and take prompt action to address them. It's essential to provide unique and valuable content to your website visitors while ensuring that search engines can correctly index and rank your pages.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Seeking Advice on Improving On-Page SEO for my Website
I'm looking for some expert advice on improving the on-page SEO for my website, CCTV Camera Installation. Despite following best practices, I feel that my site isn't performing as well as it could be in search engine rankings. Here are a few specific areas where I'd appreciate some guidance: Content Optimization: I've ensured my content is keyword-rich and valuable to my audience. Are there any advanced techniques or tools that can help further optimize my content? Meta Tags and Descriptions: I've written unique meta titles and descriptions for each page, but I'm unsure if they're as effective as they could be. What are some tips for crafting compelling meta tags that improve click-through rates? Internal Linking: I've set up a basic internal linking structure. How can I enhance this to better distribute link equity and improve user navigation? Page Load Speed: While my site's load speed is decent, I'm aware that even minor improvements can have a significant impact on SEO. Are there any specific optimizations or tools you recommend for speeding up my website? Technical SEO: I've covered the basics like XML sitemaps and robots.txt files. Are there any advanced technical SEO practices that I should be aware of? Here is the link to my website for your reference: https://www.acssllc.ae/ Thank you in advance for your help! Best regards,
On-Page Optimization | | Htdbf
Israr Khan0 -
Understanding Redirects and Canonical Tags in SEO: A Complex Case
Hi everyone, nothing serious here, i'm just playing around doing my experiments 🙂
Technical SEO | | chueneke
but if any1 of you guys understand this chaos and what was the issue here, i'd appreciate if you try to explain it to me. I had a page "Linkaufbau" on my website at https://chriseo.de/linkaufbau. My .htaccess file contains only basic SEO stuff: # removed ".html" using htaccess RewriteCond %{THE_REQUEST} ^GET\ (.*)\.html\ HTTP RewriteRule (.*)\.html$ $1 [R=301,L] # internally added .html if necessary RewriteCond %{REQUEST_FILENAME}.html -f RewriteCond %{REQUEST_URI} !/$ RewriteRule (.*) $1\.html [L] # removed "index" from directory index pages RewriteRule (.*)/index$ $1/ [R=301,L] # removed trailing "/" if not a directory RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_URI} /$ RewriteRule (.*)/ $1 [R=301,L] # Here’s the first redirect: RedirectPermanent /index / My first three questions: Why do I need this rule? Why must this rule be at the top? Why isn't this handled by mod_rewrite? Now to the interesting part: I moved the Linkaufbau page to the SEO folder: https://chriseo.de/seo/linkaufbau and set up the redirect accordingly: RedirectPermanent /linkaufbau /seo/linkaufbau.html I deleted the old /linkaufbau page. I requested indexing for /seo/linkaufbau in the Google Search Console. Once the page was indexed, I set a canonical to the old URL: <link rel="canonical" href="https://chriseo.de/linkaufbau"> Then I resubmitted the sitemap and requested indexing for /seo/linkaufbau again, even though it was already indexed. Due to the canonical tag, the page quickly disappeared. I then requested indexing for /linkaufbau and /linkaufbau.html in GSC (the old, deleted page). After two days, both URLs were back in the serps:: https://chriseo.de/linkaufbau https://chriseo.de/linkaufbau.html this is the new page /seo/linkaufbau
b14ee095-5c03-40d5-b7fc-57d47cf66e3b-grafik.png This is the old page /linkaufbau
242d5bfd-af7c-4bed-9887-c12a29837d77-grafik.png Both URLs are now in the search results and all rankings are significantly better than before for keywords like: organic linkbuilding linkaufbau kosten linkaufbau service natürlicher linkaufbau hochwertiger linkaufbau organische backlinks linkaufbau strategie linkaufbau agentur Interestingly, both URLs (with and without .html) redirect to the new URL https://chriseo.de/seo/linkaufbau, which in turn has a canonical pointing to https://chriseo.de/linkaufbau (without .html). In the SERPs, when https://chriseo.de/linkaufbau is shown, my new, updated snippet is displayed. When /linkaufbau.html is shown, it displays the old, deleted page that had already disappeared from the index. I have now removed the canonical tag. I don't fully understand the process of what happened and why. If anyone has any ideas, I would be very grateful. Best regards,
Chris0 -
How to stop /tag creating duplicate content - Wordpress
Hi, I keep getting alert for duplicate content. It seems Wordpress is creating it through a /tag https://www.curveball-media.co.uk/tag/cipr/ https://www.curveball-media.co.uk/tag/pr-agencies/ Something in the way we've got Wordpress set up?
Technical SEO | | curveballmedia0 -
Pages with duplicate meta descriptions
We have around 17 pages have underscores in the URL. From the 17 pages, we have changed 3 pages URL for example if the url is test_sample_demo.html, we have changed as test-sample-demo.html After the updates, we have made redirect as follows Redirect 301 test_sample_demo.html test-sample-demo.html Presently google webmaster tool shows as "Pages with duplicate meta descriptions" & "Pages with duplicate title tags" for changed pages How to fix this. Please help us
Technical SEO | | Intellect0 -
Page disappeared from Google index. Google cache shows page is being redirected.
My URL is: http://shop.nordstrom.com/c/converse Hi. The week before last, my top Converse page went missing from the Google index. When I "fetch as Googlebot" I am able to get the page and "submit" it to the index. I have done this several times and still cannot get the page to show up. When I look at the Google cache of the page, it comes up with a different page. http://webcache.googleusercontent.com/search?q=cache:http://shop.nordstrom.com/c/converse shows: http://shop.nordstrom.com/c/pop-in-olivia-kim Back story: As far as I know we have never redirected the Converse page to the Pop-In page. However the reverse may be true. We ran a Converse based Pop-In campaign but that used the Converse page and not the regular Pop-In page. Though the page comes back with a 200 status, it looks like Google thinks the page is being redirected. We were ranking #4 for "converse" - monthly searches = 550,000. My SEO traffic for the page has tanked since it has gone missing. Any help would be much appreciated. Stephan
Technical SEO | | shop.nordstrom0 -
Why does our page show a description in english in google spain?
Hi! We have a multilingual page and I have set in Google Webmaster Tools the language preference for the root domain to be none, Spanish for the .com/es, English for the .com/en, and German for the .com/de. The title and description show in the right language in Google Germany and google UK, but in google.es (Spain) the title and description appear in English instead of Spanish. Does anybody know why could this be happening and how to fix it? kJtF3.png
Technical SEO | | inmonova0 -
Getting Google to index new pages
I have a site, called SiteB that has 200 pages of new, unique content. I made a table of contents (TOC) page on SiteB that points to about 50 pages of SiteB content. I would like to get SiteB's TOC page crawled and indexed by Google, as well as all the pages it points to. I submitted the TOC to Pingler 24 hours ago and from the logs I see the Googlebot visited the TOC page but it did not crawl any of the 50 pages that are linked to from the TOC. I do not have a robots.txt file on SiteB. There are no robot meta tags (nofollow, noindex). There are no 'rel=nofollow' attributes on the links. Why would Google crawl the TOC (when I Pinglered it) but not crawl any of the links on that page? One other fact, and I don't know if this matters, but SiteB lives on a subdomain and the URLs contain numbers, like this: http://subdomain.domain.com/category/34404 Yes, I know that the number part is suboptimal from an SEO point of view. I'm working on that, too. But first wanted to figure out why Google isn't crawling the TOC. The site is new and so hasn't been penalized by Google. Thanks for any ideas...
Technical SEO | | scanlin0