Http and https protocols being indexed for e-commerce website
-
Hi team,
Our new e-commerce website has launched and I've noticed both http and https protocols are being indexed.
Our old website was http with only the necessary pages running https (cart, checkout etc). No https pages were indexed and you couldn't access a https page if you manually typed it into the browser.
We outrank our competition by a mile, so I'm treading carefully here and don't want to undo the progress we made on the old site, so I have a few questions:
1. How exactly do we remove one protocol from the index? We are running on Drupal. We tried a hard redirect from https to http and excluded the relevant pages (cart, login etc from the redirect), but found that you could still access https pages if you we're in the cart (https) and then pressed back on the browser button for example. At that point you could browse the entire site on https.
2. Is the safer option to emulate what we had in place on the old website e.g http with only the necessary pages being https, rather than making the switch to sitewide https?
I've been struggling with this one, so any help would be much appreciated.
Jake S
-
Just checked my GA data and you're right. Referral data from mountainjade.co.nz is there. Thanks for the heads up.
I've decided to make the switch to https, so will be organising that with dev in the coming few weeks. I'll keep you posted!
Cheers for the help again Logan,
I owe ya.
-
Great!
I've decided to make the full switch to https now, rather than wait to do it.
I will report back and let you know how it all goes!
Thanks for your help Laura.
-
I don't know why this didn't cross my mind until now, but having both versions can also mess up your Google Analytics data. Going from one to the other (can't remember which direction) creates a new session. You've probably got a lot of self-referring traffic showing up in your reports.
-
Hey Bas,
My developers share your sentiment!
Both versions of the website can be accessed by both the customer and the bots, but because we use relative urls, it can switch between http and https is a single session. This is one example:
1. Land on the homepage from a google search (http homepage is indexed).
2. Browse site on http. Add something to cart. Go to cart.
3. Cart switches to https. Navigate out of cart back into website.
4. Now urls are all https because the links on our site are relative and don't specify a protocol (e.g customer is in cart and then wants to check contact us page, it's link when clicked is as follows [Contact](/contact us). So it pulls the https protocol as there is not http protocol specified in that contact us link.
Hmmm, it definitely could be effecting UX and conversion.
-
Ideally, you'll migrate the entire site to https, and Cyrus' guide is a good one. Google has some helpful info for an http to https migration at https://support.google.com/webmasters/answer/6073543?hl=en.
The canonical tag solution is for the situation where you can't or don't want to go ahead and switch the whole site over to https right away. Either way, make sure Google knows, either through 301-redirects or canonical tags, that the http and https versions are the same page.
-
Hi Laura,
Wow, when I said we have self referencing canonicals in place (through Drupal Yoast) I hadn't even thought that it could be applying a canonical to the https version of the site aswell.
I just crawled both http and https and as you're right, the following is happening:
http://example.com is canonicalized to http://example.com
https://example.com is canonicalized to https://example.com
But I'm a little confused. In my first post I was looking for help because google was indexing both http and https pages. Are you saying that it's because of these canonicals that google is indexing both? Would it index both even if I didn't have the canonicals in place but still had SSL?
Just to confirm, canonicalizing the http URLs to the https URLs will tell google to fold the http URLs into the https and only index the https version of the site? Would I need to follow the https migration guide by Cyrus when doing this, or is this not really a 'migration' to https as we're not forcing the customer to browse in https?
Bear with me!
-
I agree with the others. I think you should pick a horse and ride it. Indecision is only causing more confusion on Google's part and is going to hurt you in the long run. Google says they prefer HTTPS and I've seen evidence of that. You're already paying for an SSL so you might as well use it to the max.
As Laura said, if you've got self-referring canonical tags on both secure and non-secure URLs, you're setting yourself up for some pretty big issues.
-
Hi Jacob,
I understand the issue. I think that this way you're not making a decision where you really should:
Either you use non-ssl or either you use ssl. To continue with the both is a terrible situation: nobody really knows what the they are supposed to know.For instance: is it possible that someone starts on the thomepage (non-ssl), goes to a product page (ssl) and then to the shopping cart which is again non-ssl? If that is the case you should really check your conversion rate because that in itself might be very damaging as well.
Yours,
Bas -
When you say you currently have self referencing canonicals, is the following happening?
The page http://example.com is canonicalized to http://example.com.
The page https://example.com is canonicalized to https://example.com.
If so, this is the bigger problem because Google sees these as 2 different URLs and may index both of them. Furthermore, you could be splitting backlinks between 2 URLs unnecessarily. This duplicate issue may be part of the reason you saw organic traffic drop when you launched your new site.
If the HTTPS URLs are already being indexed by Google, go ahead and canonicalize the http URLs to the https URLs. In other words, http://example.com will canonicalize to https://example.com.
By setting up the canonical this way, Google will fold the two URLs together and correctly treat them as the same page.
-
Good morning Laura,
Thanks for the advice.
I've replied below to Logan giving a little context. If you could take a look and let me know your thoughts it would be a huge help.
-
Hi again Logan,
I've tossed up whether or not to make the full switch to https for a while now. I'll give you a little background so you understand my position:
When our new website launched, our organic search traffic took a dip of around 15%. It has taken around two months for it to recover (almost). We changed site structure out of necessity but followed best practise to ensure we didn't undo alot of the work we had done with the old website. With the 15% organic rankings dip we saw a corresponding dip in revenue, so what I don't want to do is muddy the waters anymore than they already are by adding more moving parts to the mix (migration / redesign / http to https). And we cannot risk another dip in revenue so close to the first which may come with a full https migration (do you think?).
This is why I'm leaning toward replicating what we had in place on the old website and only forcing https on the necessary pages.
Now that you understand my position, would you still recommend the switch to https? I would love to know your thoughts.
The catch with all of this is I'm not sure exactly how the http https was implemented on the old website. At that point in time I had no need to know.
We currently have self referencing canonicals which you know we need to maintain, particularly on product pages which use URL parameters. We are also using relative links across the entire website.
Therefore, what would be the best solution here? Down the rabbit hole we go...
Thanks for your time,
-
Hi Jacob,
Cyrus Shepard put together a great guide on HTTPS migrations. Since you've already got an SSL, you may as well apply it to the whole site and set your preferred domain as HTTPS (as Laura and Bas mentioned). In the guide, he details the best ways to ensure search engines index the version you want via 301 redirect rules, canonical tags, and XML sitemaps. Don't forget to set up Search Console properties for HTTPS - www and non-www versions and set your preferred domain there as well.
Run this query in Google to monitor what they've got in their index as the canonical domain: info:mountainjade.co.nz
-
Agree with Laura: better to let the https be indexed. Nice links by the way for this topic.
Bas
-
In your case, the best thing to do is set up canonical tags to let Google know which version of the URL should be indexed. That way, it doesn't matter if Google can access the https page, and you won't have the duplicate content problem that you have now.
I can't advise you on the best way to set this up with Drupal, but you'll need to be wary of any type of automatic canonical tags. You may end up with an "http" canonical link on the http page and an "https" canonical link on the https page. That doesn't solve the problem at all.
If you are not already familiar with canonical tags, you can learn more at the links below.
- https://support.google.com/webmasters/answer/139066?hl=en
- https://moz.com/learn/seo/canonicalization
- https://webmasters.googleblog.com/2013/04/5-common-mistakes-with-relcanonical.html
By the way, I would set it up so that Google indexes the https version of your pages rather than the http version.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
New Website SEO Implications
Hi Moz Community, A client of mine has launched a new website. The new website is well designed, mobile friendly, fast loading and offers a far better UX than the old site. It has similar content but 'less wordy'. The old website was tired, slow, not mobile responsive etc but still ranked well. The domain has marketing leading authority and link metrics. Since the launch, the rankings for virtually every word has plummeted. Even previously ranked #1 words have disappeared to page 3 or 4. New pages have different URLs (301s from the old urls are working fine) and still score the same 98% (using the Moz page optimiser tool). Is it usual to experience some short term pain, or are these rankings drop an indication that something else is missing? My theory is that the new URLs are being treated like new pages, and that those new pages don't have the engagement data which is used for ranking. Thus, despite having the same authority of the old pages, as far as user data is concerned, they are new pages and therefor, not ranking well - yet. That theory would make logical sense but I'm hoping some experts here can help. Any suggestions welcome. Here's a quick checklist of things I have already done: complete 301 redirect list
Intermediate & Advanced SEO | | I.AM.Strategist
New sitemap
Submitted to console
Created internal links from within their large blog
Optimised all the new pages (img alts, H1s etc) Extra info: Platform changed from Wordpress to Expression engine
Target pages now on level 3 not level 2 (extra subfolder used)
Less words used (average word count per page from 400+ to 250) Thanks in advance 🙂0 -
Need help in de-indexing URL parameters in my website.
Hi, Need some help.
Intermediate & Advanced SEO | | ImranZafar
So this is my website _https://www.memeraki.com/ _
If you hover over any of the products, there's a quick view option..that opens up a popup window of that product
That popup is triggered by this URL. _https://www.memeraki.com/products/never-alone?view=quick _
In the URL you can see the parameters "view=quick" which is infact responsible for the pop-up. The problem is that the google and even your Moz crawler is picking up this URL as a separate webpage, hence, resulting in crawl issues, like missing tags.
I've already used the webmaster tools to block the "view" parameter URLs in my website from indexing but it's not fixing the issue
Can someone please provide some insights as to how I can fix this?0 -
Problems with a website-help
Soooooo, I did a crawl report on this site : www.greatwesternflooring.com and this was what was on the report. This is a dnn site. I'm guessing the site has a redirect loop given the http status code. Can anyone help me with a fix. (the developers have said there is no redirect on the site......clearly there is....) | http://www.greatwesternflooring.com/ | 2015-01-07T21:32:25Z | 609 : Redirect to already-visited URL received for page request. | Error attempting to request page; see title for details. | 302 | http://www.greatwesternflooring.com | <colgroup><col width="319"> <col width="144"> <col width="378"> <col span="39" width="64"></colgroup>
Intermediate & Advanced SEO | | Britewave
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | |0 -
My website is not indexing
Hello Experts As i search site :http://www.louisvuittonhandbagss.com or just entering http://www.louisvuittonhandbagss.com on Google i am not getting my website . I have done following steps 1. I have submitted sitemaps and indexed all the site maps 2.i have used GWT feature fetch as Google . 3. I have submitted my website to top social book marking websites and to some classified sites also . Pleae
Intermediate & Advanced SEO | | aschauhan5210 -
Google is Really Slow to Index my New Website
(Sorry for my english!) A quick background: I had a website at thewebhostinghero.com which had been slapped left and right by Google (both Panda & Penguin). It also had a manual penalty for unnatural links which had been lifted in late april / early may this year. I also had another domain, webhostinghero.com, which was redirecting to thewebhostinghero.com. When I realized I would be better off starting a new website than trying to salvage thewebhostinghero.com, I removed the redirection from webhostinghero.com and started building a new website. I waited about 5 or 6 weeks before putting any content on webhostinghero.com so Google had time to notice that the domain wasn't redirecting anymore. So about a month ago, I launched http://www.webhostinghero.com with 100% new content but I left thewebhostinghero.com online because it still brings a little (necessary) income. There are no links between the websites except on one page (www.thewebhostinghero.com/speed/) which is set to "noindex,nofollow" and is disallowed to search engines in robots.txt. I made sure the web page was deindexed before adding a "nofollow" link from thewebhostinghero.com/speed => webhostinghero.com/speed Since the new website launch, I've been publishing new content (from 2 to 5 posts) daily. It's getting some traction from social networks but it gets barely any clicks from Google search. It seems to take at least a week before Google indexes new posts and not all posts are indexed. The cached copy of the homepage is 12 days old. In Google Webmaster Tools, it looks like Google isn't getting the latest sitemap version unless I resubmit it manually. It's always 4 or 5 days old. So is my website just too young or could it have some kind of penalty related to the old website? The domain has 4 or 5 really old spammy links from the previous domain owner which I couldn't get rid of but otherwise I don't think there's anything tragic.
Intermediate & Advanced SEO | | sbrault740 -
Backlinks from one website to my 3 websites (hosted in 1 c-block) ?
We are making some linkbuilding. And have very nice backlinks offer. So we are planning to put our 3 websites in it. Our 3 websites are on separate IP, but same C-block. Can it be a red flag for google? Can i put my 3 backlinks in one blog post?
Intermediate & Advanced SEO | | bele0 -
Indexing non-indexed content and Google crawlers
On a news website we have a system where articles are given a publish date which is often in the future. The articles were showing up in Google before the publish date despite us not being able to find them linked from anywhere on the website. I've added a 'noindex' meta tag to articles that shouldn't be live until a future date. When the date comes for them to appear on the website, the noindex disappears. Is anyone aware of any issues doing this - say Google crawls a page that is noindex, then 2 hours later it finds out it should now be indexed? Should it still appear in Google search, News etc. as normal, as a new page? Thanks. 🙂
Intermediate & Advanced SEO | | Alex-Harford0 -
Removing a Page From Google index
We accidentally generated some pages on our site that ended up getting indexed by google. We have corrected the issue on the site and we 404 all of those pages. Should we manually delete the extra pages from Google's index or should we just let Google figure out that they are 404'd? What the best practice here?
Intermediate & Advanced SEO | | dbuckles0