Geolocation and Indexing
-
Hi all,
Our company owns site that have over 5 millions pages in Google index. We are locating in German, but our business aimed to US market.
So, recently I checked index of our site using region targeting in US and there were only 150k of pages, but when I checked targeting in German there were almost 5 billion pages.
Our server/IP locating in US, all the backlinks are from US sites.
So, why there it is only small part of the site indexed in US?
Regards,
Dmitry
-
What exact do you mean by "when I checked targeting in German there were almost 5 billion pages"? I know you are referring to "million" but how did you arrive at that number?
Other things to check:
-
submit an updated sitemap to Google. How many pages show in the site map?
-
what type of navigation does your site offer? Is all of the navigation visible in HTML?
-
some sites offer dozens of versions of the same page. A print-friendly version, sorted ascending by price, sort descending by price, sort by size and many other properties. Each sort is a different page on your site. You can have a site with 150k worth of canonical pages, but 5 million actual pages. Google will not list the duplicate pages.
-
-
Sorry, that was my mistake, I meant 5 million pages.
Unfortunately I can't name the domain name. Our site is 100% US based, with English content. I'm asking for is there some other issues (not Panda, content, etc.), that can cause the situation with regional indexation.
-
Our company owns site that have over 5 billion pages in Google index.
In order to help you, some specifics would be needed. What is the URL of the site?
Off the top of my head I would think Amazon.com is one of the biggest sites around and they have around 320 million pages indexed. The largest forum site in the world has about 16 million pages indexed by Google.
The only site I can think of with billions of indexed pages would be a scraper or other form of content manipulation website.
You mentioned you are located in Germany so clearly your pages are going to be considered most relevant there. If you wish to be more relevant to US sites, the content would need to be presented in US English, use English measurements, currency, references, etc. You would desire links from US sites as well. You could go into Google WMT and set US as your preferred country, but that would mean you would lose a significant amount of your German indexing.
Also consider the US has fully implemented Panda. It is coming to Germany but has not been implemented there yet (to the best of my knowledge, I could be mistaken). If you have a billion or more pages, I am going to speculate a huge percentage of pages are duplicated both internally to your site, and externally to the internet. If that is the case, the number of indexed pages will take a huge hit.
If your site is deemed untrustworthy due to scraped content, your entire site may be de-indexed until the issue is resolved.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Problem to get multilingual posts indexed on Google
Last year on June I decided to make my site multi-lingual. The domain is: https://www.dailyblogprofits.com/ The main language English and I added Portuguese and a few posts on Spanish. What happened since then? I started losing traffic from Google and posts on Portuguese are not being indexed. I use WPML plugin to make it multi-lingual and I had Yoast installed. This week I uninstalled Yoast and when I type on google "site:site:dailyblogprofits.com/pt-br" I started seeing Google indexing images, but still not the missing posts. I have around 145 posts on Portuguese, but on Search Console it show only 57 hreflang tags. Any idea what is the problem? I'm willing to pay for an SEO Expert to resolve this problem to me.
International SEO | | Cleber0090 -
Google does not index UK version of our site, and serves US version instead. Do I need to remove hreflanguage for US?
Webmaster tools indicates that only 25% of pages on our UK domain with GBP prices is indexed.
International SEO | | lcourse
We have another US domain with identical content but USD prices which is indexed fine. When I search in google for site:mydomain I see that most of my pages seem to appear, but then in the rich snippets google shows USD prices instead of the GBP prices which we publish on this page (USD price is not published on the page and I tested with an US proxy and US price is nowhere in the source code). Then I clicked on the result in google to see cached version of page and google shows me as cached version of the UK product page the US product page. I use the following hreflang code: rel="alternate" hreflang="en-US" href="https://www.domain.com/product" />
rel="alternate" hreflang="en-GB" href="https://www.domain.co.uk/product" /> canonical of UK page is correctly referring to UK page. Any ideas? Do I need to remove the hreflang for en-US to get the UK domain properly indexed in google?0 -
Low Index: 72 pages submitted and only 1 Indexed?
Hi Mozers, I'm pretty stuck on this and wondering if anybody else can give me some heads up around what might be causing the issues. I have 3 top level domains, NZ, AU, and USA. For some od reason I seem to be having a real issue with these pages indexing and also the sitemaps and I'm considering hiring someone to get the issue sorted as myself or my developer can''t seem to find the issues. I have attached an example of the sitemap_au.xml file. As you can see there is only 1 page that has been indexed and 72 were submitted. Basically because we host all of our domains on the same server, I was told last time our sitemaps were possibly been overwritten hence the reason why we have sitemap_au.xml and its the same for the other sitemap_nz.xml and sitemap_us.xml I also orignially had sitemap.xml for each. Another issue I am having is the meta tag des for each home page in USA and AU are showing the meta tag for New Zealand but when you look into the com and com.au code meta tag description they are all different as you can see here http://bit.ly/1KTbWg0 and here http://bit.ly/1AU0f5k Any advice around this would be so much appreciated! Thanks Justin new
International SEO | | edward-may0 -
International Sites - Sitemaps, Robots & Geolocating in WMT
Hi Guys, I have a site that has now been launched in the US having originally just been UK. In order to accommodate this, the website has been set-up using directories for each country. Example: domain.com/en-gb domain.com/en-us As the site was originally set-up for UK, the sitemap, robots file & Webmaster Tools account were added to the main domain. Example: domain.com/sitemap.xml domain.com/robots.txt The question is does this now need changing to make it specific for each country. Example: The sitemap and robots.txt for the UK would move to: domain.com/en-gb/sitemap.xml domain.com/en-gb/robots.txt and the US would have its own separate sitemap and robots.txt. Example : domain.com/en-us/sitemap.xml domain.com/en-us/robots.txt Also in order to Geolocate this in WMT would this need to be done for each directory version instead of the main domain? Currently the WMT account for the UK site is verified at www.domain.com, would this need reverifying at domain.com/en-gb? Any help would be appreciated! Thanks!
International SEO | | CarlWint0 -
Understanding the "Index Status" Data Inside Google Webmaster Tools
Currently there are total 2,787 Articles added to my Blog. The Index Status shows the following report under Index Status>Advance Total Indexed = 12,505 Blocked by robots = 8,659 And when I do search for site:techmaish.com in Google.com, it shows; About 12,200 results (0.15 seconds) Now my question. 1:- Is it normal Or there is something wrong? 2:- If there is something wrong then what is that? Thanks in advance. _ Attached is the screenshot of my GWT._ 7dk.png
International SEO | | techmaish0 -
Massive jump in pages indexed (and I do mean massive)
Hello mozzers, I have been working in SEO for a number of years but never seen anything like a jump in pages indexed of this proportion (image is from the Index Status report in Google Webmaster Tools: http://i.imgur.com/79mW6Jl.png Has anyone has ever seen anything like this?
International SEO | | Lina-iWeb
Anyone have an idea about what happened? One thing that sprung to mind might be that the same pages are now getting indexed in several more google country sites (e.g. google.ca, google.co.uk, google.es, google.com.mx) but I don't know if the Index Status report in WMT works like that. A few notes to explain the context: It's an eCommerce website with service pages and around 9 different pages listing products. The site is small - only around 100 pages across three languages 1.5 months ago we migrated from three language subdomains to a single sub-domain with language directories. Before and after the migration I used hreflang tags across the board. We saw about 50% uplift in traffic from unbranded organic terms after the migration (although on day one it was more like +300%), especially from more language diversity. I had an issue where the 'sort' links on the product tables were giving rise to thousands of pages of duplicate content, although I had used the URL parameter handling to communicate to Google that these were not significantly different and only to index the representative URL. About 2 weeks ago I blocked them using the robots.txt (Disallow: *?sort). I never felt these were doing us too much harm in reality although many of them are indexed and can be found with a site:xxx.com search. At the same time as adding *?sort to the robots.txt, I made an hreflang sitemap for each language, and linked to them from an index sitemap and added these to WMT. I added some country specific alternate URLs as well as language just to see if I started getting more traffic from those countries (e.g. xxx.com/es/ for Spanish, xxx.com/es/ for Spain, xxx.xom/es/ for Mexico etc). I dodn't seem to get any benefit from this. Webmaster tools profile is for a URL that is the root domain xxx.com. We have a lot of other subdomains, including a blog that is far bigger than our main site. But looking at the Search Queries report, all the pages listed are on the core website so I don't think it is the blog pages etc. I have seen a couple of good days in terms of unbranded organic search referrals - no spike or drop off but a couple of good days in keeping with recent improvements in these kinds of referrals. We have some software mirror sub domains that are duplicated across two website: xxx.mirror.xxx.com and xxx.mirror.xxx.ca. Many of these don't even have sections and Google seemed to be handling the duplication, always preferring to show the .com URL despite no cross-site canonicals in place. Very interesting, I'm sure you will agree! THANKS FOR READING! 79mW6Jl.png0 -
Robots.txt issue with indexation
Hello i have a problem with one of the rules for robots.txt i have a multilingual mutation of entire page on www.example.com/en/ I want to make indexable /allow/ the main page under /en/ but not indexable /disallow/ everything else under /en/* Please help me how to write the rule.
International SEO | | profesia0 -
Geolocation Questions
I'm looking to combine my company's US web presence and its United Kingdoms web presence under one common look-feel and company name. Seeing as how we are fairly small, I'm thinking the best way to do this would be to simply create a "uk" folder and creating UK specific content in there. I would also like to have some geolocation on the site to make sure users receive the content that is relevant to them. With that in mind, here my questions: 1. Would creating a "locations" page with links between the UK and the US versions of the site, be enough so that Google is sure to crawl all content? (As I understand it, Google would appear as an American visitor to my geolocation script, and wouldn't see UK content unless there was a page that would explicitly direct it in that direction, correct?) 2. I've read elsewhere that I can target specific folders to a specific geographic target using Google Webmaster Tools. However, if the "main" site is US specific (there would not be a "us" folder) Setting the geographic target for JUST the "uk" folder would still work? 3. Finally, there will unfortunately be some duplicate content between the two sites. (we have a catalog of courses, for example, that contain different groupings of courses between the two sites, but the individual courses will appear with the same descriptions within the sites) What would be the best way to deal with something like that? I would hate to point all canonical links back to the US "main" site on every instance of duplicates, but I'm not sure how else to deal with it? Thanks for any help you can give. I know this is all a bit top level, but I'm a bit paralyzed with fear of starting, seeing as how I've never had to deal with these questions before...
International SEO | | TroyCarlson0