Can anyone help me understand why google is "Not Selecting" a large number of my webpages to include when crawling my site.
-
When looking through my google webmaster tools, I clicked into the advanced settings under index status and was surprised to see that google has marked around 90% of my pages on my site as "Not Selected" when crawling. Please take a look and offer any suggestions.
-
Thank you. Thank you. Thank you. That makes so much sense. This is also the issue I am having with my communities and cities pages, pointing at my http://luxuryhomehunt.com/homes-for-sale page.
Does that make sense?
-
Thanks for the response. The pop up is running in java, and from what I have been told search engines can crawl pages so long as the opt in is running in java. Typically a visitor would hit one of our landing pages such as http://luxuryhomehunt.com/homes-for-sale/Longwood/alaqua-lakes.html where they can find information about the specific community they are searching for then if they click on a listing they would be prompted to opt in.
Do you think there may be any correlation to me using or not using canonical tags? Another thing I was wondering is if it has anything to do with, my handling of pages 2,3,4,5 etc of a city or community with more than ten listings.
I am not sure as to why your connection would have been refused, I am currently running a xml sitemap generator and maybe that had something to do with it. Either way, I am super grateful for your help and for you looking at this. I am very new to SEO and trying to learn my way through as much as possible.
-
Hmm, I just tried to click on a listing in Google but I was served a popup which required that I enter in my contact information before I could access the site http://luxuryhomehunt.com/view-property/40096215. Did you just add this pop up? Since there is no way for users to opt out of entering in contact information to view a listing, then it may be possible that the search engines are being blocked as well.
I also tried crawling the site with Screaming Frog SEO Spider and Xenu, but my connection was refused... not sure if my IP was blocked or if the site is blocking crawlers, but my guess is the search engines may be having some trouble accessing all of the pages on your site.
At the very least, I'd recommend removing that popup since it's bad for user experience and may be causing problems with the search engines.
EDIT - I did some more digging and looked the Google cache for one of your listings - http://webcache.googleusercontent.com/search?q=cache:L6LzTqj9gQUJ:luxuryhomehunt.com/view-property/40445850+&cd=6&hl=en&ct=clnk&gl=us&client=firefox-a. On this page, you have the rel="canonical" tag set to http://luxuryhomehunt.com/view-property so that tells the search engines that all of your property listing pages should use that canonical URL, which explains why most of your pages are "Not Selected" per Google -
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=2642366
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=139066
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Site Audit Tools Not Picking Up Content Nor Does Google Cache
Hi Guys, Got a site I am working with on the Wix platform. However site audit tools such as Screaming Frog, Ryte and even Moz's onpage crawler show the pages having no content, despite them having 200 words+. Fetching the site as Google clearly shows the rendered page with content, however when I look at the Google cached pages, they also show just blank pages. I have had issues with nofollow, noindex on here, but it shows the meta tags correct, just 0 content. What would you look to diagnose? I am guessing some rogue JS but why wasn't this picked up on the "fetch as Google".
Technical SEO | | nezona0 -
Can we re rank our Penalyzed website in Google?
Hello This is Maqbul, from India. I have a jobs portal blog [ bharatrecruit.com]. It was getting around 50K to 100K Views a Day and made me $100 a day. But after a few months, my competitor made negative SEO with 12,000 Spammy backlinks. Suddenly my site was hit by Google and now it is getting 200 to 300 Pageviews a day. So the question is I did not disavow bad links for a long time like 3 to 4 months. Now I disavow all the bad links but the website is not ranking. Can we re-rank this site or create another website. Please reply must. None of the bloggers can answer this. Thanks, Regards Maqbul
Technical SEO | | vinaso960 -
I hope someone can help me with page indexing problem
I have a problem with all video pages on www.tadibrothers.com.
Technical SEO | | TadiBrothers
I can not understand why google do not index all the video pages?
I never blocked them with the robots.txt file, there are no noindex/nofollow tags on the pages. The only video page that I found in search results is the main video category page: https://www.tadibrothers.com/videos and 1 video page out of 150 videos: https://www.tadibrothers.com/video/front-side-rear-view-cameras-for-backup-camera-systems I hope someone can point me to the right way0 -
Google not crawling the website from 22nd October
Hi, This is Suresh. I made changes to my website and I see that google is unable to crawl my website from 22nd October. Even it is not showing any content when I use Cache:www.vonexpy.com. Can any body help me in knowing why Google is unable to crawl my website. Is there any technical issue with the website? Website is www.vonexpy.com Thanks in advance.
Technical SEO | | sureshchowdary1 -
Does using data-href="" work more effectively than href="" rel="nofollow"?
I've been looking at some bigger enterprise sites and noticed some of them used HTML like this: <a <="" span="">data-href="http://www.otherodmain.com/" class="nofollow" rel="nofollow" target="_blank"></a> <a <="" span="">Instead of a regular href="" Does using data-href and some javascript help with shaping internal links, rather than just using a strict nofollow?</a>
Technical SEO | | JDatSB0 -
Massive Nonsensical 301 on Large ecommerce Site
We are in the process of launching a large ecommerce site, which is a rebuild. Their old URL structure does not make it possible in our eyes to logically map every URL to it's corresponding new page. We have done our best to properly and manually redirect all pages that were receiving any amount of organic traffic and have also covered all pages that had external links. Our question is we will end up with potentially tens of thousands of 404 errors that will never fix themselves. The manual work will need to stop at some point. Would it be better to leave these 404's the way they are and just let them fall out of the index or should we take everything we cannot assign appropriately to a page like the products root or the home page? I'm also open to hearing any suggestions about how others have solved massive nonsensical 301's. Thanks in advance,
Technical SEO | | Bevelwise0 -
Google having trouble accessing my site
Hi google is having problem accessing my site. each day it is bringing up access denied errors and when i have checked what this means i have the following Access denied errors In general, Google discovers content by following links from one page to another. To crawl a page, Googlebot must be able to access it. If you’re seeing unexpected Access Denied errors, it may be for the following reasons: Googlebot couldn’t access a URL on your site because your site requires users to log in to view all or some of your content. (Tip: You can get around this by removing this requirement for user-agent Googlebot.) Your robots.txt file is blocking Google from accessing your whole site or individual URLs or directories. Test that your robots.txt is working as expected. The Test robots.txt tool lets you see exactly how Googlebot will interpret the contents of your robots.txt file. The Google user-agent is Googlebot. (How to verify that a user-agent really is Googlebot.) The Fetch as Google tool helps you understand exactly how your site appears to Googlebot. This can be very useful when troubleshooting problems with your site's content or discoverability in search results. Your server requires users to authenticate using a proxy, or your hosting provider may be blocking Google from accessing your site. Now i have contacted my hosting company who said there is not a problem but said to read the following page http://www.tmdhosting.com/kb/technical-questions/other/robots-txt-file-to-improve-the-way-search-bots-crawl/ i have read it and as far as i can see i have my file set up right which is listed below. they said if i still have problems then i need to contact google. can anyone please give me advice on what to do. the errors are responce code 403 User-agent: *
Technical SEO | | ClaireH-184886
Disallow: /administrator/
Disallow: /cache/
Disallow: /components/
Disallow: /includes/
Disallow: /installation/
Disallow: /language/
Disallow: /libraries/
Disallow: /media/
Disallow: /modules/
Disallow: /plugins/
Disallow: /templates/
Disallow: /tmp/
Disallow: /xmlrpc/0 -
My report only says it crawled 1 page of my site.
My report used to crawl my entire site which is around 90 pages. Any idea of why this would happen? www.treelifedesigns.com
Technical SEO | | nathan.marcarelli0