Moving from http to https: image duplicate issue?
-
Hello everyone,
We have recently moved our entire website virtualsheetmusic.com from http:// to https:// and now we are facing a question about images.
Here is the deal: All webpages URLs are properly redirected to their corresponding https if they are called from former http links. Whereas, due to compatibility issues, all images URLs can be called either via http or https, so that any of the following URLs work without any redirect:
http://www.virtualsheetmusic.com/images/icons/ResponsiveLogo.png
https://www.virtualsheetmusic.com/images/icons/ResponsiveLogo.png
Please note though that all internal links are relative and not absolute.
So, my question is: Can that be a problem from the SEO stand point? In particular: We have thousands of images indexed on Google, mostly images related to our digital sheet music preview image files, and many of them are ranking pretty well in the image pack search results. Could this change be detrimental in some way? Or doesn't make any difference in the eyes of Google? As I wrote above, all internal links are relative, so an image tag like this one:
Hasn't changed at all, it is just loaded in a https context.
I'll wait for your thoughts on this. Thank you in advance!
-
No problem
-
Great! Glad to know that. Thank you Dimitrii, I appreciated your help very much!
-
Oh, I see. Yeah, there shouldn't be any problems, if someone else links to your images with http. And yes, your assumption is correct
-
Thank you Dimitrii to clarifying, actually all our webpages now load images only via the https://, but since many external websites are hard-linking to many of our images via the regular http:// protocol, I was thinking to allow linking to them the "insecure" way if requested. Do you see my point? So... to better clarify my initial question, let's say Google is spidering one of those external affiliates and finds an image tag like this:
Will Google consider the image found at:
http://www.virtualsheetmusic.com/image.jpg
a duplicate of:
https://www.virtualsheetmusic.com/image.jpg
?? This was my original question...
In any case, I have made some testings today, and I have been able to redirect all images via .htaccess permanently (301) to https:// and looks like even if an image is requested with the http:// from the browser, it shows up correctly because the web browser handles redirects for images in the same way it handles them for the web page itself.
So... my concern should be solved this way. But in case, for any reason, I need to be able to serve the same image from both protocols (http or https) it is my understand that that shouldn't be an issue anyway. Is my assumption correct?
Thanks again.
-
I did quick search, and there are lots of good articles about why images are not duplicate content:Â http://bfy.tw/9Qy4
-
So, the reason I recommend having images loading only through one resource is the "insecurity" of https connection, if any resources are loaded not over https. You might have seen that sometimes instead of green lock in a browser bar, it can show yellow exclamation mark - that's one of the reasons. And also it's just cleaner, if everything is loaded the same way.
Here is a link to resource about mixed content:Â https://developers.google.com/web/fundamentals/security/prevent-mixed-content/fixing-mixed-content
-
Thank you Dimitrii for your reply.
Well, your two statements above contradicts each other, in my opinion. You see, what really concerns me is your last suggestion:
"it's better to make sure that images (and all the other resources) available only through one protocol - http or https."
And hence my original concern. Why should we make sure that images are available only through one protocol if you say first that there isn't such thing as duplicate content for images? Why should we concern about that then?
Sorry for my further request for clarification. I really appreciated your help!
-
Howdy.
As far as I understand, there is no such thing as duplicate content just for images. Duplicate content is more for the page as a whole. Especially, since you guys redirected all the links, you shouldn't have any problems, since google will simply "realize" the change.
Now, it's better to make sure that images (and all the other resources) available only through one protocol - http or https.
Hope this helps
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Main menu duplication
I am working on a site that has just gone through a migration to Shopify at the very same time as Google did an update in October. So problems from day 1. All main menu categories have subsequently over the past 6 weeks fallen off a cliff. All aspects of the site have been reviewed in terms of technical, link profile and on-page, with the site in better shape than several ranking competitors. One issue that i'd like some feedback on is the main menu which has 4 iterations in the source. desktop desktop (sticky) mobile mobile (sticky - appears as a second desktop sticky but I assume for mobile) These items that are "duplicated" menus are the top level menu items only. The rest of the nested menu items are included within the last mobile menu option. So desktop menu in source doesn't include any of the sub-menu items, the mobile version carries all these there are 4 versions of the top level main menu items in source Should I be concerned? Considering we have significant issues should this be cleaned up?
Intermediate & Advanced SEO | | MickEdwards0 -
Infinite scrolling issue?
Hi Guys, Reviewing this E-commerce page -Â https://tinyurl.com/ybjjwr65 Based on this Google article:Â https://webmasters.googleblog.com/2014/02/infinite-scroll-search-friendly.html It mentions: Make sure that you or your content management system produces a paginated series (component pages) to go along with your infinite scroll. How would you check this, is there a tool to conduct this test? Cheers.
Intermediate & Advanced SEO | | kayl870 -
Hreflang implementation issue
We are currently handling search for a global brand www.example.com which has presence in many countries worldwide. To help Google understand that there is an alternate version of the website available in another language, we have used “hreflang” tags. Also, there is a mother website (www.example.com/global) which is given the attribution of “x-default” in the “hreflang” tag. For Malaysia as a geolocation, the mother website is ranking instead of the local website (www.example.com/my) for majority of the products. The code used for “hreflang” tag execution, on a product page, being: These “hreflang” tags are also present in the XML sitemap of the website, mentioning them below: <loc>http://www.example.com/my/product_name</loc> <lastmod>2017-06-20</lastmod> Is this implementation of “hreflang” tags fine? As this implementation is true across all geo-locations, but the mother website is out-ranking me only in the Malaysia market. If the implementation is correct, what could be other reasons for the same ranking issue, as all other SEO elements have been thoroughly verified and they seem fine.
Intermediate & Advanced SEO | | Starcom_Search0 -
No images in Google index
No images are indexed on this site (client of ours):Â http://www.rubbermagazijn.nl/. We've tried everything (descriptive alt texts, image sitemaps, fetch&render, check robots) but a site:www.rubbermagazijn.nl shows 0 image results and the sitemap report in Search Console shows 0 images indexed. We're not sure how to proceed from here. Is there anyone with an idea what the problem could be?
Intermediate & Advanced SEO | | Adriaan.Multiply0 -
Have a Robots.txt Issue
I have a robots.txt file error that is causing me loads of headaches and is making my website fall off the SE grid. on MOZ and other sites its saying that I blocked all websites from finding it. Could it be as simple as I created a new website and forgot to re-create a robots.txt file for the new site or it was trying to find the old one? I just created a new one. Google's website still shows in the search console that there are severe health issues found in the property and that it is the robots.txt is blocking important pages. Does this take time to refresh? Is there something I'm missing that someone here in the MOZ community could help me with?
Intermediate & Advanced SEO | | primemediaconsultants0 -
Image Maps
Hey forum, I'm curious about Image Maps. Few things I'm not sure about: 1. Will the links be followed? If so, will Google respect rel="nofollow"? 2. Will the image be considered 1 image? (indexed as image, etc.) Or will each map segment be treated as a separate image? 3. Any other SEO pros\cons to consider when adding an image map to an existing page? Thanks, Corwin.
Intermediate & Advanced SEO | | corwin0 -
Duplicate content mess
One website I'm working with keeps a HTML archive of content from various magazines they publish. Some articles were repeated across different magazines, sometimes up to 5 times. These articles were also used as content elsewhere on the same website, resulting in up to 10 duplicates of the same article on one website. With regards to the 5 that are duplicates but not contained in the magazine, I can delete (resulting in 404) all but the highest value of each (most don't have any external links). There are hundreds of occurrences of this and it seems unfeasible to 301 or noindex them. After seeing how their system works I can canonical the remaining duplicate that isn't contained in the magazine to the corresponding original magazine version - but I can't canonical any of the other versions in the magazines to the original. I can't delete the other duplicates as they're part of the content of a particular issue of a magazine. The best thing I can think of doing is adding a link in the magazine duplicates to the original article, something along the lines of "This article originally appeared in...", though I get the impression the client wouldn't want to reveal that they used to share so much content across different magazines. The duplicate pages across the different magazines do differ slightly as a result of the different Contents menu for each magazine. Do you think it's a case of what I'm doing will be better than how it was, or is there something further I can do? Is adding the links enough? Thanks. 🙂
Intermediate & Advanced SEO | | Alex-Harford0 -
Login Page = Duplicate content?
I am having a problem with duplicate content with my log in page QuickLearn Online Anytime - Log-in
Intermediate & Advanced SEO | | QuickLearnTraining
http://www.quicklearn.com/maven/login.aspx
QuickLearn Online Anytime - Log-in
http://www.quicklearn.com/maven/login.aspx?ReturnUrl=/maven/purchase.aspx?id=BAM-SP
QuickLearn Online Anytime - Log-in
http://www.quicklearn.com/maven/login.aspx?ReturnUrl=/maven/purchase.aspx?id=BRE-SP
QuickLearn Online Anytime - Log-in
http://www.quicklearn.com/maven/login.aspx?ReturnUrl=/maven/purchase.aspx?id=BTAF
QuickLearn Online Anytime - Log-in
http://www.quicklearn.com/maven/login.aspx?ReturnUrl=/maven/purchase.aspx?id=BTDF What is the best way to handle it? Add a couple sentences to each page to make it unique? Use a rel canonical, or a no index no follow or something completely different? Your help is greatly appreciated!0