Geo-location by state/store
-
Hi there,
We are a Grocery co-operative retailer and have chain of stores owned by different people. We are building a new website, where we would geo-locate the closest store to the customer and direct them to a particular store (selected based on cookie and geo location). All our stores have a consistent range of products + Variation in 25% range. I have few questions
-
How to build a site-map. Since it will be mandatory for a store to be selected and same flow for the bot and user, should have all products across all stores in the sitemap? we are allowing users to find any products across all stores if they search by product identifier. But, they will be able to see products available in a particular store if go through the hierarchical journey of the website.
-
Will the bot crawl all pages across all the stores or since it will be geolocated to only one store, the content belonging to only one store will be indexed?
-
We are also allowing customers to search for older products which they might have bought few years and that are not part of out catalogue any more. these products will not appear on the online hierarchical journey but, customers will be able to search and find the products . Will this affect our SEO ranking?
Any help will be greatly appreciated.
Thanks - Costa
-
-
If you consistently see the IP address and redirect, or change content, based only on that then you will want to exempt Googlebot from those personalizations in one way or another. There are many options to this, like blocking the resources that handle this (i.e. the JavaScript.js file associated with personalization based on history or geo-location), or what was suggested above. Blocking that piece of script in the robots.txt file is less likely to be seen as cloaking.
All of this begs the question though: If you're looking at the IP, then setting a cookie, then updating the content based on the cookie, it shouldn't be an issue in the first place. Googlebot isn't accepting your cookies. So if I were to browse in Incognito mode using Chrome (and thus not accept cookies), would I see the same site and product assortments no matter which location I was in? If that's the case, maybe you don't have a problem. This is pretty easy to test.
Ultimately, I think you're going to want a single product page for each Sku, rather than one for each product at each location. The content, pricing, etc.. can be updated by location if they have a cookie, but the URL should probably never change - and the content shouldn't change by IP if they don't have a cookie.
1. Check IP
2. Embed their location in a cookie
3. Set cookie
4. If cookie is excepted and thus exists, do personalize.
If the cookie does not exist, do not personalize. You can show a message that says you must accept cookies to get the best experience, but don't make it block any major portion of the content.
-
Thanks for this. Few clarifications please,
Isnt having a different journey for a user and bot cloaking? Will google not penalise a site for that? - To make it clear - we have a single website and based on the Geo of the user, we will filter product availability. If a customer is from state A, we will should "X" products and if a customer is from State B, we will show X+Y or X-Y. All the products will have a canonical URL as part of the sitemap, so even if the product is not navigatable through the hierarchy on the website, crawlers will be able to find it through the direct canonical URL.
Here us a link to the article where John Mueller from google has some comments on the subject - https://www.seroundtable.com/google-geolocation-redirects-are-okay-26933.html
I have picked excerpts from you reply where I have some doubts, great if you can throw more light into these?
-
- "It seems like you'll have to have the same products available in multiple stores. You will want them all indexed, but will have to work hard to differentiate them (different images, different copy, different Meta data) otherwise Google will probably pick one product from one store as 'canonical' and not index the rest, leading to unfair product purchasing (users only purchasing X product from Y store, never the others)"
Since, we will have same (X products) across all our stores and across stores these products will have a single canonical URL, what will be the advantage of having different content by stores. we are thinking the content on the product pages will be the same, but, the availability of the product alone will differ based on geo. The sitemap will also remain the same across stores with the canonical product URLs
-
- "Will the bot crawl all pages across all the stores or since it will be geolocated to only one store, the content belonging to only one store will be indexed?" - No it won't. Every time Google crawls from a different data centre, they will think all your other pages are being redirected now and that part of the site is now closed. Exempt Googlebot's user-agent from your redirects or face Google's fiery wrath when they fail to index anything properly
Could you please explain a bit more on what do you mean by re-direct, as all products will exists in the website for a crawler to see if the canonical URL is used for crawling. Only the availability and the product visibility through the navigation journey will change based on geo.
Thank you for your time on this. Its extremely useful
Thanks - Costa
-
-
-
"We are a Grocery co-operative retailer and have chain of stores owned by different people. We are building a new website, where we would geo-locate the closest store to the customer and direct them to a particular store (selected based on cookie and geo location). All our stores have a consistent range of products + Variation in 25% range. I have few questions" - make sure you exempt Googlebot's user-agent from your geo-based redirects otherwise the crawling of your site will end up in a big horrible mess
-
"How to build a site-map. Since it will be mandatory for a store to be selected and same flow for the bot and user, should have all products across all stores in the sitemap? we are allowing users to find any products across all stores if they search by product identifier. But, they will be able to see products available in a particular store if go through the hierarchical journey of the website." - any pages you want Google to index should be in your XML sitemap. Any pages you don't want Google ti index should not be in there (period). If a URL uses a canonical tag to point somewhere else (and thus marks itself as NON-canonical) it shouldn't be in the XML sitemap. If a URL is blocked via robots.txt or Meta no-index directives, it shouldn't be in the XML sitemap. If a URL results in an error or redirect, it shouldn't be in your XML sitemap.The main thing to concern yourself with, is creating a 'seamless' view of indexation for Google. It seems like you'll have to have the same products available in multiple stores. You will want them all indexed, but will have to work hard to differentiate them (different images, different copy, different Meta data) otherwise Google will probably pick one product from one store as 'canonical' and not index the rest, leading to unfair product purchasing (users only purchasing X product from Y store, never the others). In reality, setting out to build a site which such highly divergent duplication is never going to yield great results, you'll just have to be aware of that from the outset
-
"Will the bot crawl all pages across all the stores or since it will be geolocated to only one store, the content belonging to only one store will be indexed?" - No it won't. Every time Google crawls from a different data centre, they will think all your other pages are being redirected now and that part of the site is now closed. Exempt Googlebot's user-agent from your redirects or face Google's fiery wrath when they fail to index anything properly
-
"We are also allowing customers to search for older products which they might have bought few years and that are not part of out catalogue any more. these products will not appear on the online hierarchical journey but, customers will be able to search and find the products . Will this affect our SEO ranking?" - If the pages are orphaned except in the XML sitemap, their rankings will go down over time. It won't necessarily hurt the rest of your site, though. Sometimes crappy results are better than no results at all!
Hope that helps
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
I have followed all the steps in google speed ranking on how to increase my website http://briefwatch.com/ speed but no good result
My website http://briefwatch.com/ has a very low-speed score on google page speed and I followed all the steps given to me still my website speed doesn't increase
Local Website Optimization | | Briefwatch0 -
Expanding to Other Geo Locations
Our company originally started in one city, now it is in multiple and the city we started in is actually now less important to our business than some of the new cities. We've of course have Google Places for Business listings for all our cities and are listed in the other prominent directories for each City (Bing Places, Manta, Superpages, etc etc) We have created once city page for each city in our domain. All this has improved our Local SERPs for those cities but they pale in comparison to our dominance in the city we started out in. We did have the first city in our home page title, we took that out. The obvious problem is from an SEO standpoint your home page is your "strongest" page but how do you make your home page rank top for multiple location intent searches: "{city} {target keyword}" ? The 3-pack is KEY. For example, for one city we make it into the local 3-pack but we are not in the organic SERPs on page 1 outside of the 3-pack. As far as I can tell the major factor in 3-pack ranking is of course proximity of business to the user's location or user's location intent. I would say followed by the natural ranking factors (or at least a large subset of them) that Google uses for its normal organic rankings, followed by Google Places reviews. You would think the Google places reviews really make a difference, but not as much as you think. So how do you dominate local searches in different cities when competing against local-only companies? My only guess is you need to create as much content as possible. You don't want to make micro sites I think as you lose all the link juice going to your main site. But how much content can one make that isn't duplicative. You can describe the same products and services over and over for each city but that's not useful nor wise. I guess you could do some re-writing. But other than a different address, phone, and staff members, if your service is identical for each city it doesn't leave a lot of room for useful content creation to improve local search SERPS. I guess this begs the overall question, can a multi-city company ever dominate local SERPS when the search has a location intent (city name in the search) it there is even just a couple competing local companies doing some SEO work. it seems it is an extreme uphill battle if not next to impossible. (Never say never.)
Local Website Optimization | | Searchout1 -
I have a USA Shopify store and want to create a UK store. Should I use a domain or a sub-domain for the new store?
I'm currently running 2x Shopify Stores (names changed from actual): "coolpants.com" is our UK store "us.coolpants.com" is our USA store However, Google Analytics has shown that most of the traffic to "coolpants.com" comes from the USA, and this market has our biggest growth potential. So we want to switch our domain setup to the below: Make "coolpants.com" our USA store Stop using "us.coolpants.com" and redirect it to "coolpants.com" Set up our UK store as a new domain ("coolpants.co.uk) or subdomain ("uk.coolpants.com") My question is – should I set up the UK store as a domain or a sub-domain? What are the pros and cons of each? Thank you in advance for any help!
Local Website Optimization | | benkapow0 -
In local SEO, how important is it to include city, state, and state abbreviation in doctitle?
I'm trying to balance local geographic keywords with product keywords. I appreciate the feedback from the group! Michael
Local Website Optimization | | BFMichael0 -
What is the effect of CloudFlare CDN on page load speeds, hosting IP location and the ultimate SEO effect?
Will using a CDN like CloudFlare.com confuse search engines in terms of the location (IP address) of where the site is actually physically hosted especially since CloudFlare distributes the site's content all around the globe? I understand it is important that if customers are mostly in a particular city it makes sense to host on an IP address in the same city for better rankings, all things else being equal? I have a number of city-based sites but does it make having multiple hosting plans in multiple cities/ countries (to be close to customers) become suddenly a ridiculous thing with a CDN? In other words should I just reduce it down to having one hosting plan anywhere and just use the CDN to distribute it? I am really struggling with this concept trying to understand if I should consolidate all my hosting plans under one, or if I should get rid of CloudFlare entirely (can it cause latency in come cases) and create even more locally-based hosting plans (like under site5.com who allow many city hosting plans). I really hope you can help me somehow or point me to an expert who can clarify this confusing conundrum. Of course my overall goal is to have:
Local Website Optimization | | uworlds
1. lowest page load times
2. best UX
3. best rankings I do realise that other concepts are more important for rankings (great content, and links etc.) but assuming that is already in place and every other factor is equal, how can I fine tune the hosting to achieve the desirable goals above? Many thanks!
Mark0 -
Local SEO - Adding the location to the URL
Hi there, My client has a product URL: www.company.com/product. They are only serving one state in the US. The existing URL is ranking in a position between 8-15 at the moment for local searches. Would it be interesting to add the location to the URL in order to get a higher position or is it dangerous as we have our rankings at the moment. Is it really giving you an advantage that is worth the risk? Thank you for your opinions!
Local Website Optimization | | WeAreDigital_BE
Sander0 -
Location Pages and Duplicate Content and Doorway Pages, Oh My!
Google has this page on location pages. It's very useful but it doesn't say anything about handling the duplicate content a location page might have. Seeing as the loctions may have very similar services. Lets say they have example.com/location/boston, example.com/location/chicago, or maybe boston.example.com or chicago.example.com etc. They are landing pages for each location, housing that locations contact information as well as serving as a landing page for that location. Showing the same services/products as every other location. This information may also live on the main domains homepage or services page as well. My initial reaction agrees with this article: http://moz.com/blog/local-landing-pages-guide - but I'm really asking what does Google expect? Does this location pages guide from Google tell us we don't really have to make sure each of those location pages are unique? Sometimes creating "unique" location pages feels like you're creating **doorway pages - **"Multiple pages on your site with similar content designed to rank for specific queries like city or state names". In a nutshell, Google's Guidelines seem to have a conflict on this topic: Location Pages: "Have each location's or branch's information accessible on separate webpages"
Local Website Optimization | | eyeflow
Doorway Pages: "Multiple pages on your site with similar content designed to rank for specific queries like city or state names"
Duplicate Content: "If you have many pages that are similar, consider expanding each page or consolidating the pages into one." Now you could avoid making it a doorway page or a duplicate content page if you just put the location information on a page. Each page would then have a unique address, phone number, email, contact name, etc. But then the page would technically be in violation of this page: Thin Pages: "One of the most important steps in improving your site's ranking in Google search results is to ensure that it contains plenty of rich information that includes relevant keywords, used appropriately, that indicate the subject matter of your content." ...starting to feel like I'm in a Google Guidelines Paradox! Do you think this guide from Google means that duplicate content on these pages is acceptable as long as you use that markup? Or do you have another opinion?0 -
Killing it in Yahoo/Bing...Sucking it in Google. What gives?
Our website http://www.survive-a-storm.com has historically performed well in Google for the search terms "storm shelters" and "tornado shelters." Our geographic focus is nationwide, but we are particularly interested in ranking up for Oklahoma. Right now we are hovering at about the third position in Yahoo/Bing, and in some geographic areas (i.e., as selected in Google's search settings) we are doing reasonably to quite well for these terms in Google (i.e., first page). In Oklahoma, though, we are holding steady around positions 20-25. We have just changed the title tag on our home page, cleaned up a bit of on-page optimization, and are going to work on getting some more optimized content on the page. We are outperforming the competition on Domain Authority (38) and Page Authority (46), and as far as I can tell, other key metrics are respectable. Our social isn't bad, but could always use improvement--which we are working on. Any idea why we might be lagging so badly in Google? Any help would be appreciated!
Local Website Optimization | | Survive-a-Storm0