Google Not Indexing XML Sitemap Images
-
Hi Mozzers,
We are having an issue with our XML sitemap images not being indexed.
The site has over 39,000 pages and 17,500 images submitted in GWT. If you take a look at the attached screenshot, 'GWT Images - Not Indexed', you can see that the majority of the pages are being indexed - but none of the images are.
The first thing you should know about the images is that they are hosted on a content delivery network (CDN), rather than on the site itself. However, Google advice suggests hosting on a CDN is fine - see second screenshot, 'Google CDN Advice'. That advice says to either (i) ensure the hosting site is verified in GWT or (ii) submit in robots.txt. As we can't verify the hosting site in GWT, we had opted to submit via robots.txt.
There are 3 sitemap indexes: 1) http://www.greenplantswap.co.uk/sitemap_index.xml, 2) http://www.greenplantswap.co.uk/sitemap/plant_genera/listings.xml and 3) http://www.greenplantswap.co.uk/sitemap/plant_genera/plants.xml.
Each sitemap index is split up into often hundreds or thousands of smaller XML sitemaps. This is necessary due to the size of the site and how we have decided to pull URLs in. Essentially, if we did it another way, it may have involved some of the sitemaps being massive and thus taking upwards of a minute to load.
To give you an idea of what is being submitted to Google in one of the sitemaps, please see view-source:http://www.greenplantswap.co.uk/sitemap/plant_genera/4/listings.xml?page=1.
Originally, the images were SSL, so we decided to reverted to non-SSL URLs as that was an easy change. But over a week later, that seems to have had no impact. The image URLs are ugly... but should this prevent them from being indexed?
The strange thing is that a very small number of images have been indexed - see http://goo.gl/P8GMn. I don't know if this is an anomaly or whether it suggests no issue with how the images have been set up - thus, there may be another issue.
Sorry for the long message but I would be extremely grateful for any insight into this. I have tried to offer as much information as I can, however please do let me know if this is not enough.
Thank you for taking the time to read and help.
Regards,
Mark
-
Hi Mark,
I'm just following the thread as I have a similar problem. Would you mind sharing your results from the tests?
Thanks,
Bogdan -
Thanks Everett - that's exactly what I intend to do.
We will be testing two new sitemaps with 100 x URLs each. 1. With just the file extension removed and 2. With the entire cropping part of the URL removed, as suggested by Matt.
Will be interested to see whether just one or both of the sitemaps are successful. Will of course post the outcome here, for anyone who might have this problem in future.
-
It isn't always that simple. Maybe commas don't present a problem on their own. Maybe double file extensions don't present a problem on their own. Maybe a CDN doesn't present a problem on its own. Maybe very long, complicated URLs don't present a problem on their own.
You have all of these. Together, in any combination, they could make indexation of your images a problem for Google.
Just test it out on a few. Get rid of the file extension. If that doesn't work, get rid of the comma. That is all you can do. Start with whatever is easiest for the developer to implement, and test it out on a few before rolling it out across all of your images.
-
Cheers for that mate - especially the useful Excel formula.
I am going to try a few things in isolation so that we can accurately say which element/s caused the issue.
Thanks again, mate.
-
Ignore the developer - what worked for one doesn't mean it'll work for you
The easiest way to test this is to manually create a sitemap with 100 or so 'clean' image URLs. Just pull the messy ones into excel and use the formula below to create a clean version (Use A1 for messy, B1 for formula).
Good luck mate.
=CONCATENATE("image:imageimage:lochttp://res.cloudinary.com/greenplantswap/image/upload/",RIGHT(A1,LEN(A1)-(FIND("",(SUBSTITUTE(A1,"/","",(IF(LEN(TRIM(A1))=0,0,LEN(TRIM(A1))-LEN(SUBSTITUTE(A1,"/",""))))))))),"</image:loc></image:image>")
-
Thanks for the responses guys, much appreciated.
In terms of the commas, that was something that I put to the developer, however he was able to come back with examples where this has clearly not been an issue - e.g. apartable.com have commas in their URLs and use the same CDN (Coudinary).
However, I agree with you that double file extension could be the issue. I may have to wait until next week to find out as the developer is working on another project, but will post the outcome here once I know.
Thank you again for the help!
-
Hello Edlondon,
I think you're probably answering your own question here. Google typically doesn't have any problem indexing images served from a CDN. However, I've seen Google have problems with commas in the URL at times. Typically it happens when other elements in the URL are also troublesome, such as your double file extension.
Are you able to rename the files to get rid of the superfluous .jpg extension? If so, I'd recommend trying it out on a few dozen images. We could come up with a lot of hypothesis, but that would be the one I'd test first.
-
Hmmm I step off here, never used cloudinary.com or even heard of them. I personally use NetDNA, with pull zones (which means that they load the image/css/js from your origin and store a version on their servers) while handling cropping/resizing from my own end (via PHP and then loading that image, example: http://cdn.fulltraffic.net/blog/thumb/58x58/youtube-video-xQmQeKU25zg.jpg try changing the 58x58 to another size and my server will handle the crop/resize while NetDNA will serve it and store for future loads).
-
Found one of the sites with the same Cloudinary URLs with commas - apartable.com
See Google image results: https://www.google.co.uk/search?q=site:apartable.com&tbm=isch
Their images appear to be well indexed. One thing I have noticed, however, is that we often have .jpg twice in the image URL. E.g.:
- http://res.cloudinary.com/greenplantswap/image/upload/c_crop,g_north,h_0.9,w_1.0/c_fill,d_no_image_icon-720x720**.jpg**,g_center,h_900,q_80,w_900/v1352574983/oyfos82vwvmxdx91hxaw**.jpg**
- http://res.cloudinary.com/greenplantswap/image/upload/c_crop,g_north,h_0.9,w_1.0/c_fill,d_no_image_icon-720x720**.jpg**,g_center,h_900,q_80,w_900/v1352574989/s09cv3krfn7gbyvw3r2y**.jpg**
- http://res.cloudinary.com/greenplantswap/image/upload/c_crop,g_north,h_0.9,w_1.0/c_fill,d_no_image_icon-720x720**.jpg**,g_center,h_407,q_80,w_407/v1352575010/rl7cl4xi0timza1sgzxj**.jpg**
Wonder if that is confusing Google? If so, none of this is consistent, as they do have a few images indexed with exactly the same kind of URL as those listed above.
-
Thought I had them on email but must be within our fairly cumbersome Skype thread... let me have a dig through when I get chance and I'll post them up here.
-
Hmmmm, okay... Could you post the examples they gave, and an example page where the images are located on the site?
-
Hi Matt,
Thought I should let you know that (i) the X-Robots-Tag was not set, so that's not the issue and (ii) the URLs, although ugly, are not the issue either. We had a couple of examples of websites with the same thing (I'm told the commas facilitate on-the-fly sizing and cropping) and their images were indexed fine.
So, back to the drawing board for me! Thank you very much for the suggestions, really do appreciate it.
Mark
-
Hmm interesting - we hadn't thought of the X-Robots-Tag http header. I'm going to fire that over to the developer now.
As for the URLs, they are awful! But I am told that this is not a problem - but perhaps this is worth re-chasing up as other solutions have, so far, been unfruitful.
Thanks for taking the time to help, Matt - I'll let you know if that fixes it! Unfortunately it could be another week before I know, as the developer is currently working on another project so any changes may be early-mid next week.
Thanks again...
-
This is a bit of a long shot but if the files have been uploaded using their API it may have been that the 'X-Robots-Tag' http header is set to no-index...
Also, those URLs don't look great with the commas in them. Have you tried doing a small subset that just has the image id (e.g. http://res.cloudinary.com/greenplantswap/image/upload/nprvu0z6ri227cgnpmqc.jpg)?
Matt
-
Hi Federico,
Thanks very much for taking the time to respond.
To answer your question, we are using http://cloudinary.com/. So, taking one of the examples from the XML sitemap I posted above, an example of an image URL is http://res.cloudinary.com/greenplantswap/image/upload/c_crop,g_north,h_0.9,w_1.0/c_fill,d_no_image_icon-720x720.jpg,g_center,h_900,q_80,w_900/v1352575097/nprvu0z6ri227cgnpmqc.jpg (what a lovely URL!).
I had a look at http://res.cloudinary.com/robots.txt and it seems that they are not blocking anything - the disallow instruction is commented out. I assume that is indeed the robots.txt I should be looking at?
Assuming it is, this does not appear to get to the bottom of why the images are not being indexed.
Any further assistance would be greatly appreciated - we have 17k unique images that could be driving traffic and this is a key way that people find our kind of website.
Thanks,
Mark
-
Within that robot.txt file on the CDN (which one are you using?) have you set to allow Google to index them?
Most CDNs I know allows you to block engines via the robots.txt to avoid bandwidth consumption.
In the case you are using NetDNA (MaxCDN) or the like, make sure your robots file isn't disallowing robots to crawl.
We are using a CDN too to deliver images and static files and all of them are being indexed, we tested disallowing crawlers but it caused a lot of warnings, so instead we no allow all of them to read and index content (is a small price to pay to have your content indexed).
Hope that helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
No images in Google index
No images are indexed on this site (client of ours): http://www.rubbermagazijn.nl/. We've tried everything (descriptive alt texts, image sitemaps, fetch&render, check robots) but a site:www.rubbermagazijn.nl shows 0 image results and the sitemap report in Search Console shows 0 images indexed. We're not sure how to proceed from here. Is there anyone with an idea what the problem could be?
Intermediate & Advanced SEO | | Adriaan.Multiply0 -
Any excellent recommendations for a sitemap.xml plugin?
Hi, I'm trying to find a sitemap generator/plugin that I can point my client to. My client is using Magento, and is one of the largest sports store i Norway (around 20 000 products). I've heard there's one that can set the <priority>according to page views, sold units, and other relevant parameters, and that also takes care of the other elements in the sitemap.xml.</priority> Any good recommendations out there? 🙂
Intermediate & Advanced SEO | | Inevo0 -
How to update campaign product on google merchant xml feed?
Hello Guys, I am using google merchant xml which updates on every morning 12 am, now many times we run campaign which starts at 8 am and for that product price differ due to that google reject such products. So my query is, is here anyway to set campaign start and end time in xml? Note - In API we can update product price anytime but in xml it can be done at one time only that is 12 am only. right?
Intermediate & Advanced SEO | | varo0 -
XML Sitemap & Bad Code
I've been creating sitemaps with XML Sitemap Generator, and have been downloading them to edit on my pc. The sitemaps work fine when viewing in a browser, but when I download and open in Dreamweaver, the urls don't work when I cut and paste them in the Firefox URL bar. I notice the codes are different. For example, an "&" is produced like this..."&". Extra characters are inserted, producing the error. I was wondering if this is normal, because as I said, the map works fine when viewing online.
Intermediate & Advanced SEO | | alrockn0 -
Why would one of our section pages NOT be indexed by Google?
One of our higher traffic section pages is not being indexed by Google. The products that reside on this section page ARE indexed by Google and are on page 1. So why wouldn't the section page be even listed and indexed? The meta title is accurate, meta description is good. I haven't received any notices in Webmaster Tools. Is there a way to check to see if OTHER pages might also not be indexed? What should a small ecom site do to see about getting it listed? SOS in Modesto. Ron
Intermediate & Advanced SEO | | yatesandcojewelers0 -
Are links that are disavowed with Google Webmaster Tools removed from the Google Webmaster Profile for the domain?
Hi, Two part question - First, are links that you disavow using google webmaster tools ever removed from the webmaster tools account profile ? Second, when you upload a file to disavow links they ask if you'd like to replace the previously uploaded file. Does that mean if you don't replace the file with a new file that contains the previously uploaded urls those urls are no longer considered disavowed? So, should we download the previous disavow file first then append the new disavow urls to the file before uploading or should we just upload a new file that contains only the new disavow urls? Thanks
Intermediate & Advanced SEO | | bgs0 -
Is there a way to keep sitemap.xml files from getting indexed?
Wow, I should know the answer to this question. Sitemap.xml files have to be accessible to the bots for indexing they can't be disallowed in robots.txt and can't block the folder at the server level. So how can you allow the bots to crawl these xml pages but have them not show up in google's index when doing a site: command search, or is that even possible? Hmmm
Intermediate & Advanced SEO | | irvingw0 -
XML Sitemap Index Percentage (Large Sites)
Hi all I'm wanting to find out from those who have experience dealing with large sites (10s/100s of millions of pages). What's a typical (or highest) percentage of indexed pages vs. submitted pages you've seen? This information can be found in webmaster tools where Google shows you the pages submitted & indexed for each of your sitemap. I'm trying to figure out whether, The average index % out there There is a ceiling (i.e. will never reach 100%) It's possible to improve the indexing percentage further Just to give you some background, sitemap index files (according to schema.org) have been implemented to improve crawl efficiency and I'm wanting to find out other ways to improve this further. I've been thinking about looking at the URL parameters to exclude as there are hundreds (e-commerce site) to help Google improve crawl efficiency and utilise the daily crawl quote more effectively to discover pages that have not been discovered yet. However, I'm not sure yet whether this is the best path to take or I'm just flogging a dead horse if there is such a ceiling or if I'm already at the average ballpark for large sites. Any suggestions/insights would be appreciated. Thanks.
Intermediate & Advanced SEO | | danng0