Google images
-
Hi,
I am working on a website with a large number (millions) of images. For the last five months Ihave been trying to get Google Images to crawl and index these images (example page: http://bit.ly/1ePQvyd). I believe I have followed best practice in the design of the page, naming of images etc. Whilst crawlng and indexing of the pages is going reasonably well with the standard crawler, the image bot has only crawled about half a million images and indexed only about 40,000.
Can anyone suggest what I could do to increase this number 100 fold?
Richard
-
Hi Richard,
I'm not a Google Images expert (otherwise I'd offer to help!) but you could check out the companies / individuals listed on Moz's <a>Recommended Providers page</a>. Many are agencies that prefer to work with people on longer retainers, but you may find someone to work with there, or the people listed may be able to point you to a trusted contact who specialises in images.
I haven't heard of people specifically specialising in image search but that doesn't mean those people don't exist, or that you won't find someone who has a lot of experience with image-heavy photography / art websites.
Cheers,
Jane
-
Hi,
Thanks for those who have offered advice so far.
I am looking an buy the services of an SEO expert with experience in getting vast numbers of images indexed by Google Images. If that is you or you know someone who could help, get in touch.
Richard
-
Thanks,
I think we could sharpen up the title and some other things.
However, we are doing reasonably well with indexing of pages - it's the images that seem to be so slow.
We do have image sitemaps (eg http://bit.ly/RDHmyo) but our experiments show that this is not much help.
This sitemap above was submitted about 14 days ago and has had about 2 million pages indexed but only 5,000 images indexed.
It feels like I have my foot on the hose pipe
Richard
-
Thanks for looking at this.
We are adding sitemaps "slowly" - ie 10 million at a time. eg: http://bit.ly/RDHmyo
This stemap was submitted about 14 days ago and has had about 2 million pages indexed but only 5,000 images indexed.
We will add another this week.
I think you may well be right about crawl budget. I have ensured that the bots are almost exclusively focused on these pages (well over 90% of crawls are of these pages). I am assuming that the only way t increase budget is to increase authority. Is that right?
-
I don't know about a 100-fold increase, but as Martijn said, an image sitemap will certainly help. You can often build images into the standard sitemap, but with so many images, you are best keeping these separate.
Something else I would look to do, is change the page Title structure. Currently, this says:
"The sky begins to lighten during morning twilight behind a suspension tower of the Leonard P. Zakim Bunker Hill Memorial Bridge. - DYFXET | Alamy"There is no mention here of photography, royalty free, stock images, etc.
I would be cutting the description of the image down so it looked something like this:
"The sky begins to lighten during morning twilight behind a suspension tower... Royalty free stock photography | Alamy"
Look at others who do well in stock image searches (Shutterstock, 123rf, etc.) and match some of what they do.
-Andy
-
Hi Richard,
Have you tried submitting image sitemap files to both Google Webmaster Tools as Bing Webmaster Tools. It's probably related to your crawl budget that Google is not crawling the millions of images on your site. That's why they maybe also can't found more than these pages.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google Pagination Changes
What with Google recently coming out and saying they're basically ignoring paginated pages, I'm considering the link structure of our new, sooner to launch ecommerce site (moving from an old site to a new one with identical URL structure less a few 404s). Currently our new site shows 20 products per page but with this change by Google it means that any products on pages 2, 3 and so on will suffer because google treats it like an entirely separate page as opposed to an extension of the first. The way I see it I have one option: Show every product in each category on page 1. I have Lazy Load installed on our new website so it will only load the screen a user can see and as they scroll down it loads more products, but how will google interpret this? Will Google simply see all 50-300 products per category and give the site a bad page load score because it doesn't know the Lazy Load is in place? Or will it know and account for it? Is there anything I'm missing?
Intermediate & Advanced SEO | | moon-boots0 -
Javascript content not being indexed by Google
I thought Google has gotten better at picking up unique content from javascript. I'm not seeing it with our site. We rate beauty and skincare products using our algorithms. Here is an example of a product -- https://www.skinsafeproducts.com/tide-free-gentle-he-liquid-laundry-detergent-100-fl-oz When you look at the cache page (text) from google none of the core ratings (badges like fragrance free, top free and so forth) are being picked up for ranking. Any idea what we could do to have the rating incorporated in the indexation.
Intermediate & Advanced SEO | | akih0 -
Google slow to index pages
Hi We've recently had a product launch for one of our clients. Historically speaking Google has been quick to respond, i.e when the page for the product goes live it's indexed and performing for branded terms within 10 minutes (without 'Fetch and Render'). This time however, we found that it took Google over an hour to index the pages. we found initially that press coverage ranked until we were indexed. Nothing major had changed in terms of the page structure, content, internal linking etc; these were brand new pages, with new product content. Has anyone ever experienced Google having an 'off' day or being uncharacteristically slow with indexing? We do have a few ideas what could have caused this, but we were interested to see if anyone else had experienced this sort of change in Google's behaviour, either recently or previously? Thanks.
Intermediate & Advanced SEO | | punchseo0 -
Removing Parameterized URLs from Google Index
We have duplicate eCommerce websites, and we are in the process of implementing cross-domain canonicals. (We can't 301 - both sites are major brands). So far, this is working well - rankings are improving dramatically in most cases. However, what we are seeing in some cases is that Google has indexed a parameterized page for the site being canonicaled (this is the site that is getting the canonical tag - the "from" page). When this happens, both sites are being ranked, and the parameterized page appears to be blocking the canonical. The question is, how do I remove canonicaled pages from Google's index? If Google doesn't crawl the page in question, it never sees the canonical tag, and we still have duplicate content. Example: A. www.domain2.com/productname.cfm%3FclickSource%3DXSELL_PR is ranked at #35, and B. www.domain1.com/productname.cfm is ranked at #12. (yes, I know that upper case is bad. We fixed that too.) Page A has the canonical tag, but page B's rank didn't improve. I know that there are no guarantees that it will improve, but I am seeing a pattern. Page A appears to be preventing Google from passing link juice via canonical. If Google doesn't crawl Page A, it can't see the rel=canonical tag. We likely have thousands of pages like this. Any ideas? Does it make sense to block the "clicksource" parameter in GWT? That kind of scares me.
Intermediate & Advanced SEO | | AMHC0 -
Buying a domain banned by google
Hi , I came across a super domain for my business but found out that it was a great domain with 100s of link backs but is now banned by Google search engine meaning Google does not index content from that domain. Since the domains linkbacks are from my domin does it make sense to but that domain and redirect those link backs to another (301) and hope that the new domain gets some juice ... I know it is sounding crazy and may not be the best thing to do ethically but still wanted to check if its possible to get some juice.. Rgds Avinash
Intermediate & Advanced SEO | | Avinashmb0 -
What is next from Google Panda and Google Penguin?
Does anyone know what we can expect next from Google Panda/Penguin? We did prepare for this latest update and so far so good.
Intermediate & Advanced SEO | | jjgonza0 -
Why do i not receive google traffic?
over the 4-5 months i have published over 3000 unique articles which i have payed well over 10 000usd for, but i still only receive about 20 google visitors a day for that content. i uploaded the 3000 articles after i 301 redirected the old site to a a new domain (old site had 1000 articles, and at least 300visits from google a day), and all the old conetnt receives the traffic fine (301 redirect is working 100percent now and pr went from 0 to 3pr) articles are also good ranging from 400-800 words. 90 percent of them are indexed by google, most of them have been bookmarked to digg reddit etc website domain is over 10 years old - alltopics.com why google doesnt send me the traffic i deserve?
Intermediate & Advanced SEO | | rxesiv0 -
Google indexing flash content
Hi Would googles indexing of flash content count towards page content? for example I have over 7000 flash files, with 1 unique flash file per page followed by a short 2 paragraph snippet, would google count the flash as content towards the overall page? Because at the moment I've x-tagged the roberts with noindex, nofollow and no archive to prevent them from appearing in the search engines. I'm just wondering if the google bot visits and accesses the flash file it'll get the x-tag noindex, nofollow and then stop processing. I think this may be why the panda update also had an effect. thanks
Intermediate & Advanced SEO | | Flapjack0