GoogleBot Mobile & Depagination
-
I am building a new site for a client and we're discussing their inventory section. What I would like to accomplish is have all their products load on scroll (or swipe on mobile). I have seen suggestions to load all content in the background at once, and show it as they swipe, lazy loading the product images. This will work fine for the user, but what about how GoogleBot mobile crawls the page?
Will it simulate swiping? Will it load every product at once, killing page load times b/c of all of the images it must load at once? What are considered SEO best practices when loading inventory using this technique.
I worry about this b/c it's possible for 2,000+ results to be returned, and I don't want GoogleBot to try and load all those results at once (with their product thumbnail images). And I know you will say to break those products up into categories, etc. But I want the "swipe for more" experience. 99.9% of our users will click a category or filter the results, but if someone wants to swipe through all 2,000 items on the main inventory landing page, they can. I would rather have this option than "Page 1 of 350".
I like option #4 in this question, but not sure how Google will handle it.
I asked Matt Cutts to answer this, if you want to upvote this question.
https://www.google.com/moderator/#11/e=adbf4&u=CAIQwYCMnI6opfkj -
What you ideally want to do is set up the mobile site as a standard site. Then utilize javascript to call each page in an order defined by the users actions with dynamic loading.
This has two benefits:
-
SEO and SERP. The pages will be indexed as they should. If you have one huge page you are still limited to the 2 or 3 keywords as always. When you see a good infinite scroll website it is not one page, it only looks this way due to JavaScript calling additional pages at triggers that have been set.
-
No JavaScript graceful fallback (or fallforward as it is actually the native state). If you have one page, lazy loading with JavaScript and they do not support it then you have 2,000 pages worth of images loading at one time which is otherwise known as a bounce.
You will want to build out the site with no consideration to the infinite scrolling (except in design ie. tile-able backgrounds for a smooth non stop flow) then apply the script after you have a logical site structure using silo'ed categories. Google bot, Google bot mobile and users who do not have JavaScript will all have a useable site and the SERPS will rank pages as they should.
Tip: Keep any page wide bar or graphic styles in the header or the footer of the page. You will normally only call the content or article portion of the page to the infinite scroll so you have a non-stop flow on the site.
Hope this helps
I know your not using WordPress but I am assuming you are using some sort of templated PHP script for a 2K product store. This WP plugin is pretty easy to understand and what I first used to grab the concept. Also, if wanting to go a more Pinterest route look into Masonry JavaScript. http://www.infinite-scroll.com/
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Keyword Cannabalisation & Ecommerce
Hi I have an Ecommerce site, with a lot of similar products - for example leather office chairs - 80 products all very similar.. We worked to optimise product pages for longer tail phrases such as black executive leather office chair, but we now have different product pages trying to rank for these longer tail phrases as well. Now I'm trying to decide whether to focus on some priority product pages - adding lots of useful content/videos etc to try & boost the ones we want to rank for the long tail. OR whether to focus on the category page, and getting this to rank for all keyword variations... I'm a it stuck - any advice is welcome!
Intermediate & Advanced SEO | | BeckyKey0 -
Title & Keywords
Hi Quick question on arrangement of keywords in titles. I know the order isn't so important anymore, but would there be a real issue if I want to rank for 'Henry Xtra' but my title reads 'Numatic Henry Xtra Vacuum Cleaner' Rather than 'Henry Xtra Vacuum Cleaner' ?? Will it really make much difference? Thank you!
Intermediate & Advanced SEO | | BeckyKey0 -
Display:None CSS & SEO
Hi A while back I was told that using the display:none tag to hide content you want minimised is bad for onpage SEO - is this the case? It's not that we want to hide it from Google, we just don't want it taking up a huge amount of space on product pages. I have found some of these on our site, and want to know how bad they are. Is the content under the tag going to be ignored? Thank you
Intermediate & Advanced SEO | | BeckyKey0 -
What if my site isn't ready for Mobile Armageddon by April 21st??
Hello Moz Experts, I am fighting for one of our sites to be mobile optimized, but the fight is taking longer than anticipated (need approval from higher ups). What happens if my site is not ready by April 21st? Will it take long to recover, like Penguin? Or, will the recovery be fairly quick? Say I release a mobile version of my site a week later. Then Google will have to reindex it and rank me again. How long will that take before I regain my traffic? Thanks,
Intermediate & Advanced SEO | | TMI.com0 -
Difference in Number of URLS in "Crawl, Sitemaps" & "Index Status" in Webmaster Tools, NORMAL?
Greetings MOZ Community: Webmaster Tools under "Index Status" shows 850 URLs indexed for our website (www.nyc-officespace-leader.com). The number of URLs indexed jumped by around 175 around June 10th, shortly after we launched a new version of our website. No new URLs were added to the site upgrade. Under Webmaster Tools under "Crawl, Site maps", it shows 637 pages submitted and 599 indexed. Prior to June 6th there was not a significant difference in the number of pages shown between the "Index Status" and "Crawl. Site Maps". Now there is a differential of 175. The 850 URLs in "Index Status" is equal to the number of URLs in the MOZ domain crawl report I ran yesterday. Since this differential developed, ranking has declined sharply. Perhaps I am hit by the new version of Panda, but Google indexing junk pages (if that is in fact happening) could have something to do with it. Is this differential between the number of URLs shown in "Index Status" and "Crawl, Sitemaps" normal? I am attaching Images of the two screens from Webmaster Tools as well as the MOZ crawl to illustrate what has occurred. My developer seems stumped by this. He has submitted a removal request for the 175 URLs to Google, but they remain in the index. Any suggestions? Thanks,
Intermediate & Advanced SEO | | Kingalan1
Alan0 -
Disavow Links & Paid Link Removal (discussion)
Hey everyone, We've been talking about this issue a bit over the last week in our office, I wanted to extend the idea out to the Moz community and see if anyone has some additional perspective on the issue. Let me break-down the scenario: We're in the process of cleaning-up the link profile for a new client, which contains many low quality SEO-directory links placed by a previous vendor. Recently, we made a connection to a webmaster who controls a huge directory network. This person found 100+ links to our client's site on their network and wants $5/link to have them removed. Client was not hit with a manual penalty, this clean-up could be considered proactive, but an algorithmic 'penalty' is suspected based on historical keyword rankings. **The Issue: **We can pay this ninja $800+ to have him/her remove the links from his directory network, and hope it does the trick. When talking about scaling this tactic, we run into some ridiculously high numbers when you talk about providing this service to multiple clients. **The Silver Lining: **Disavow Links file. I'm curious what the effectiveness of creating this around the 100+ directory links could be, especially since the client hasn't been slapped with a manual penalty. The Debate: Is putting a disavow file together a better alternative to paying for crappy links to be removed? Are we actually solving the bad link problem by disavowing or just patching it? Would choosing not to pay ridiculous fees and submitting a disavow file for these links be considered a "good faith effort" in Google's eyes (especially considering there has been no manual penalty assessed)?
Intermediate & Advanced SEO | | Etna0 -
Re-Direct Users But Don't Affect Googlebot
This is a fairly technical question... I have a site which has 4 subdomains, all targeting a specific language. The brand owners don't want German users to see the prices on the French sub domain and are forcing users into a re-direct to the relevant subddomain, based on their IP address. If a user comes from a different country, (ie the US) they are forced on the UK sub domain. The client is insistent on keeping control of who sees what (I know that's a debate in it's own right), but these re-directs we're implementing to make that happen, are really making it difficult to get all the subdomains indexed as I think googlebot is also getting re-directed and is failing to do it's job. Is there are a way of re-directing users, but not Googlebot?
Intermediate & Advanced SEO | | eventurerob0 -
Googlebot + Meta-Refresh
Quick question, can Googlebot (or other search engines) follow meta refresh tags? Does it work anything like a 301 in terms of passing value to the new page?
Intermediate & Advanced SEO | | kchandler1