Http v https Duplicate Issues
-
Hello,
I noticed earlier an issue on my site.
http://mysite.com and https://mysite.com both had canonical links pointing to themselves so in effect creating duplicate content.
I have now taken steps to ensure the https version has a canonical that points to the http version but I was wondering what other steps would people recommend? Is it safe to NOINDEX the https pages? Or block them via robots.txt or both?
We are not quite ready to go fully HTTPS with our site yet (I know Google now prefers this)
Any thoughts would be very much appreciated.
-
Since HTTPS is now a ranking signal, it is better to use the HTTPS version as the canonical. I would personally make every page of the site HTTPS via 301 redirections (or rel=canonical but those can be trickier to implement).
http://site.com --301--> https://site.com
http://site.com/page1/ --301--> https://site.com/page1/
etc.This may require a few changes to the site (internal links shouldn't have unnecessary redirections, adding the HTTPS site to Search Consol (webmaster tools), etc.) so make sure you look around for resources on migration.
If you decide to keep HTTP only, do not noindex or disallow HTTPS because you may have valuable links pointing to HTTPS which help your ranking.
-
Thanks for your replies. Although I'm still confused.
I have areas of the site that are and should be https (checkout etc) and these pages have canonical links pointing to the https version.
The rest of my site however is still on http but the https versions can be accessed via their urls. What I have done today is to add a canonical tag to the https pages to point to the http pages. Is this the correct thing to do to avoid a duplicate content issue?
-
Hi,
I agree with Patrick, if you are not using the https then the safest way to ensure no canonical content is to remove it all together.
If you are using it partially such as checkouts and user areas, then you could 301 redirect the https traffic for the other pages to their https counterparts until you are ready to go full https.
Kind Regards
-
Hi there
If your site is not ready to go fully https, I would hold off on it until you are, unless you have a checkout process or information gathering portion of the site that should be https.
Reason being - the https isn't providing any ranking factor value as it's being canonicalized to the http version of your site, so you're not getting the value.
When you are ready to go https, I recommend taking a look at this Moz resource, specifically the section under SEO checklist to preserve your rankings.
Hope this helps! Good luck!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What is best practice for fixing urls that have duplicate content, non-static and other issues?
Hi, I know there are several good answers regarding duplicate content issues on this website already, however I have a question that involves the best way to avoid negative SEO impacts if I change the urls for an ecommerce site. Basically a new client has the following website http://www.gardenbeauty.co.uk and I notice that it suffers from duplicate content due to the http://www version and the non www version of the pages - this seems quite easy to fix using the guidance on this website. However I notice that the product page urls are far from ideal in that they have several issues including:- (a) they are mostly too long (b) don't include the keyword terms (in terms of best practice) (c) they don't use Static URLS An example of one these product urls would be http://www.gardenbeauty.co.uk/plant-details.php?name=Autumn Glory&p_genus=Hebe&code=heagl&category=hebe I'd like to address these issues, but the pages rank highly for the products themselves, therefore my question is what would you recommend I do to fix the urls without risking the high positions that many of these product pages have? thanks, Ben
Technical SEO | | bendyman0 -
XML Sitemap Issue or not?
Hi Everyone, I submitted a sitemap within the google webmaster tools and I had a warning message of 38 issues. Issue: Url blocked by robots.txt. Description: Sitemap contains urls which are blocked by robots.txt. Example: the ones that were given were urls that we don't want them to be indexed: Sitemap: www.example.org/author.xml Value: http://www.example.org/author/admin/ My issue here is that the number of URL indexed is pretty low and I know for a fact that Robot.txt aren't good especially if they block URL that needs to be indexed. Apparently the URLs that are blocked seem to be URLs that we don't to be indexed but it doesn't display all URLs that are blocked. Do you think i m having a major problem or everything is fine?What should I do? How can I fix it? FYI: Wordpress is what we use for our website Thanks
Technical SEO | | Tay19860 -
Duplicate Content Issue
Very strange issue I noticed today. In my SEOMoz Campaigns I noticed thousands of Warnings and Errors! I noticed that any page on my website ending in .php can be duplicated by adding anything you want to the end of the url, which seems to be causing these issues. Ex: Normal URL - www.example.com/testing.php Duplicate URL - www.example.com/testing.php/helloworld The duplicate URL displays the page without the images, but all the text and information is present, duplicating the Normal page. I Also found that many of my PDFs seemed to be getting duplicated burried in directories after directories, which I never ever put in place. Ex: www.example.com/catalog/pdfs/testing.pdf/pdfs/another.pdf/pdfs/more.pdfs/pdfs/ ... when the pdfs are only located in a pdfs directory! I am very confused on how to fix this problem. Maybe with some sort of redirect?
Technical SEO | | hfranz0 -
RSS Hacking Issue
Hi Checked our original rss feed - added it to Google reader and all the links go to the correct pages, but I have also set up the RSS feed in Feedburner. However, when I click on the links in Feedburner (which should go to my own website's pages) , they are all going to spam sites, even though the title of the link and excerpt are correct. This isn't a Wordpress blog rss feed either, and we are on a very secure server. Any ideas whatsoever? There is no info online anywhere and our developers haven't seen this before. Thanks
Technical SEO | | Kerry220 -
Http and Https Update
Guys, I was just wondering what is the current up-to-date resolution for HTTP and HTTPS, One of my client needs SSL for transaction,and they only need single page to be encrypted(SSLed) , So should I Force SSL that page only and direct entire HTTPS pages to HTTP to avoid duplicate page issue, or would it considered cloaking. It is been long time i did something similar, Or current Search Engines algo can handle it by themselves and i should just leave it . Thanks!
Technical SEO | | DigitalJungle0 -
Duplicate Content Issue
Hi Everyone, I ran into a problem I didn't know I had (Thanks to the seomoz tool) regarding duplicate content. my site is oxford ms homes.net and when I built the site, the web developer used php to build it. After he was done I saw that the URL's looking like this "/blake_listings.php?page=0" and I wanted them like this "/blakes-listings" He changed them with no problem and he did the same with all 300 pages or so that I have on the site. I just found using the crawl diagnostics tool that I have like 3,000 duplicate content issues. Is there an easy fix to this at all or does he have to go in and 301 Redirect EVERY SINGLE URL? Thanks for any help you can give.
Technical SEO | | blake-766240 -
Duplicate META Description
Two of my urls point to the same site. For example: 123.com and OneTwoThree.com. Because the pages have two separate url's and the same content, my Google Webmaster Tools account is showing this as Duplicate Meta Descriptions. Is there a way around this?
Technical SEO | | BradBorst0 -
Is this considered as duplicate content?
One of my clients has a template page they have used repeatedly each time they have a new news item. The template includes a two-paragraph customer quote/testimonial for the company. So, they now have 100+ pages with the same customer quote. The rest of the page content / body copy is unique. Is there any likelihood of this being considered duplicate content?
Technical SEO | | bjalc20110