Sudden Change In Indexed Pages
-
Every week I check the number of pages indexed by google using the "site:" function. I have set up a permanent redirect from all the non-www pages to www pages.
When I used to run the function for the:
non-www pages (i.e site:mysite.com), would have 12K results
www pages (i.e site:www.mysite.com) would have about 36K
The past few days, this has reversed! I get 12K for www pages, and 36K for non-www pages.
Things I have changed:
I have added canonical URL links in the header, all have www in the URL.
My questions:
Is this cause for concern?
Can anyone explain this to me?
-
Maybe Google includes all sub-domains. I just tested my site, and got the following results.
site:handsomeweb.com 340 results (all have www)
site:www.handsomeweb.com 231 results (all have www)
The difference is the first query includes pages located at:blog.handsomeweb.com
-
I don't get both resolving, and yes, I did set a preferred domain in my Google account.
Any other ideas?
-
Have you done preferred domain in Google?
I don't know how you could have done the 301's and still get both resolving. Have you run xenu or some other program to insure the 301s are there?
-
We did the 301 redirects from non-www to www when we launched the site.
I have another site that I have done a 301 from www to non-www, and you get 0 results when you search "site:www.mysite.com".
They are both on the same platform, which makes it more confusing!!!
-
inhouseseo
I have looked at several of our sites and see no change in results for site:
You stated: **Things I have changed: **
I have added canonical URL links in the header, all have www in the URL.
I believe what is happening (assuming you changed the canonical URL prior to the change in results of site:) is the change is the result of the canonical application you have added. However, I am not sure how you could still have an aggregate of 48K pages, are you sure this is accurate?
If, you are showing 12K of www, and 36K of non www, I would guess that the 12K were duplicated within the 36K. Therefore, you would have only 36K pages on your site.
Typically, when we encounter a site that has both www and non-www we select a preferred domain in WMT and do the 301 redirect in .htaccess file. Once this is done, over a short period, we will have only what we have chosen as the preferred domain www or non www.
So, if we started with 1,000 pages of www and 2,000 of non www, then if preferred choice is non www. We will end up with 2,000 pages total.
My suggestion would be to go into WMT and select a preferred domain and do the 301 redirect in the .htaccess file. Once that is done, I believe your problem will be resolved. rel=canon will not accomplish this in and of itself. Give it a few weeks and check your results.
Best
-
This is something I've been noticing greatly over the past few months. I was literally just about to post this question:
site: command, y u no accurate?!
I feel like the site:domain.com command used to be very accurate in showing you total pages indexed. Recently, I've seen wildly varied results returned.
Of course, it varies based upon the inclusion of "www.", but even without it, I've seen such results as anywhere from 193k to 8million pages... and everything in-between.
Why the variance? Has anyone else seen this recently?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How to get a large number of urls out of Google's Index when there are no pages to noindex tag?
Hi, I'm working with a site that has created a large group of urls (150,000) that have crept into Google's index. If these urls actually existed as pages, which they don't, I'd just noindex tag them and over time the number would drift down. The thing is, they created them through a complicated internal linking arrangement that adds affiliate code to the links and forwards them to the affiliate. GoogleBot would crawl a link that looks like it's to the client's same domain and wind up on Amazon or somewhere else with some affiiiate code. GoogleBot would then grab the original link on the clients domain and index it... even though the page served is on Amazon or somewhere else. Ergo, I don't have a page to noindex tag. I have to get this 150K block of cruft out of Google's index, but without actual pages to noindex tag, it's a bit of a puzzler. Any ideas? Thanks! Best... Michael P.S., All 150K urls seem to share the same url pattern... exmpledomain.com/item/... so /item/ is common to all of them, if that helps.
Intermediate & Advanced SEO | | 945010 -
Removing massive number of no index follow page that are not crawled
Hi, We have stackable filters on some of our pages (ie: ?filter1=a&filter2=b&etc.). Those stacked filters pages are "noindex, follow". They were created in order to facilitate the indexation of the item listed in them. After analysing the logs we know that the search engines do not crawl those stacked filter pages. Does blocking those pages (by loading their link in AJAX for example) would help our crawl rate or not? In order words does removing links that are already not crawled help the crawl rate of the rest of our pages? My assumption here is that SE see those links but discard them because those pages are too deep in our architecture and by removing them we would help SE focus on the rest of our page. We don't want to waste our efforts removing those links if there will be no impact. Thanks
Intermediate & Advanced SEO | | Digitics0 -
Landing pages, are my pages competing?
If I have identified a keyword which generates income and when searched in google my homepage comes up ranked second, should I still create a landing page based on that keyword or will it compete with my homepage and cause it to rank lower?
Intermediate & Advanced SEO | | The_Great_Projects0 -
Big discrepancies between pages in Google's index and pages in sitemap
Hi, I'm noticing a huge difference in the number of pages in Googles index (using 'site:' search) versus the number of pages indexed by Google in Webmaster tools. (ie 20,600 in 'site:' search vs 5,100 submitted via the dynamic sitemap.) Anyone know possible causes for this and how i can fix? It's an ecommerce site but i can't see any issues with duplicate content - they employ a very good canonical tag strategy. Could it be that Google has decided to ignore the canonical tag? Any help appreciated, Karen
Intermediate & Advanced SEO | | Digirank0 -
Thousands of Web Pages Disappered from Google Index
The site is - http://shop.riversideexports.com We checked webmaster tools, nothing strange. Then we manually resubmitted using webmaster tools about a month ago. Now only seeing about 15 pages indexed. The rest of the sites on our network are heavily indexed and ranking really well. BUT the sites that are using a sub domain are not. Could this be a sub domain issue? If so, how? If not, what is causing this? Please advise. UPDATE: What we can also share is that the site was cleared twice in it's lifetime - all pages deleted and re-generated. The first two times we had full indexing - now this site hovers at 15 results in the index. We have many other sites in the network that have very similar attributes (such as redundant or empty meta) and none have behaved this way. The broader question is how to do we get the indexing back ?
Intermediate & Advanced SEO | | suredone0 -
Can a home page penalty cause a drop in rankings for all pages?
All my main keywords have dropped out of the SERPS. Could it be that the home page (the strongest) page has been devalued and therefore 'link juice' that used to spread throughout the site is no longer doing so. Would this cause all other pages to drop? I just can't understand how all my pages have lost rankings. The site is still indexed so there's no problem there.
Intermediate & Advanced SEO | | SamCUK0 -
How do you de-index and prevent indexation of a whole domain?
I have parts of an online portal displaying in SERPs which it definitely shouldn't be. It's due to thoughtless developers but I need to have the whole portal's domain de-indexed and prevented from future indexing. I'm not too tech savvy but how is this achieved? No index? Robots? thanks
Intermediate & Advanced SEO | | Martin_S0 -
Should I prevent Google from indexing blog tag and category pages?
I am working on a website that has a regularly updated Wordpress blog and am unsure whether or not the category and tag pages should be indexable. The blog posts are often outranked by the tag and category pages and they are ultimately leaving me with a duplicate content issue. With this in mind, I assumed that the best thing to do would be to remove the tag and category pages from the index, but after speaking to someone else about the issue, I am no longer sure. I have tried researching online, but there isn't anything that provided any further information. Please can anyone with any experience of dealing with issues like this or with any knowledge of the topic help me to resolve this annoying issue. Any input will be greatly appreciated. Thanks Paul
Intermediate & Advanced SEO | | PaulRogers0