Can Googlebot crawl the content on this page?
-
Hi all,
I've read the posts in Google about Ajax and javascript (https://support.google.com/webmasters/answer/174992?hl=en) and also this post: http://moz.com/ugc/can-google-really-access-content-in-javascript-really.
I am trying to evaluate if the content on this page, http://www.vwarcher.com/CustomerReviews, is crawlable by Googlebot? It appears not to be. I perused the sitemap and don't see any ugly Ajax URLs included as Google suggests doing. Also, the page is definitely indexed, but appears the content is only indexed via its original source (Yahoo!, Citysearch, Google+, etc.).
I understand why they are using this dynamic content, because it looks nice to an end-user and requires little to no maintenance. But, is it providing them any SEO benefit? It appears to me that it would be far better to take these reviews and simply build them into HTML.
Thoughts?
-
Thanks so much Bill and Brian. This is exactly what I was thinking. I did the same thing Bill suggested initially and took a snippet from one of the reviews and did a verbatim search and got nothing. What I thought this told me was that yes, the page was indexed, but not the content. The fact that the cached version renders the content from the javascript only shows that the script was executed, not necessarily that any of the content it contains was actually indexed.
From an SEO standpoint I think this is valuable content that the dealer would very much want indexed. While the service providing the javascript might be very convenient, and the majority of end users might be able to consume the content, the fact that it's not searchable, to me, means it's an opportunity lost.
Thanks again everyone.
-
Bill is right. The page is indexed and cached, however googlebot cannot read the reviews. If you view the cache (cache:http://www.vwarcher.com/CustomerReviews) and then click "Text-only version" in the upper right, you'll see that those reviews are not there (google can't crawl them).
-
Dana,
Yes, Google has indexed the page. However, if you view the source code of the page you won't see any text of the customer reviews. Even if you view the cached version, there is text there: but it's not necessarily text on the page. If you view the cached version, and then see the source code of the page, again, the content is not there. The reviews are not there in the source code.
Let's take this one step further. Search Google for one of the testimonials or a part of one: "Marc Palermo is a great customer service" which is NOT indexed in Google.
Google sees the source code but doesn't appear to be indexing the content.
-
Hehe no problem at all Dana. Glad to have helped stand in for your coffee
-Andy
-
Hah! Thanks Andy. Must not have had enough coffee this morning. I didn't even think of looking at the cache...so obvious, lol! Thanks so much. You are spot on.
-
Google appears to have the page cached, so I would say there are no real issues.
Just do a cache:http://www.vwarcher.com/CustomerReviews and you can see what Google currently has.
-Andy
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
GoogleBot still crawling HTTP/1.1 years after website moved to HTTP/2
Whole website moved to https://www. HTTP/2 version 3 years ago. When we review log files, it is clear that - for the home page - GoogleBot continues to only access via HTTP/1.1 protocol Robots file is correct (simply allowing all and referring to https://www. sitemap Sitemap is referencing https://www. pages including homepage Hosting provider has confirmed server is correctly configured to support HTTP/2 and provided evidence of accessing via HTTP/2 working 301 redirects set up for non-secure and non-www versions of website all to https://www. version Not using a CDN or proxy GSC reports home page as correctly indexed (with https://www. version canonicalised) but does still have the non-secure version of website as the referring page in the Discovery section. GSC also reports homepage as being crawled every day or so. Totally understand it can take time to update index, but we are at a complete loss to understand why GoogleBot continues to only go through HTTP/1.1 version not 2 Possibly related issue - and of course what is causing concern - is that new pages of site seem to index and perform well in SERP ... except home page. This never makes it to page 1 (other than for brand name) despite rating multiples higher in terms of content, speed etc than other pages which still get indexed in preference to home page. Any thoughts, further tests, ideas, direction or anything will be much appreciated!
Technical SEO | | AKCAC1 -
Page content not being recognised?
I moved my website from Wix to Wordpress in May 2018. Since then, it's disappeared from Google searches. The site and pages are indexed, but no longer ranking. I've just started a Moz campaign, and most pages are being flagged as having "thin content" (50 words or less), when I know that there are 300+ words on most of the pages. Looking at the page source I find this bit of code: page contents Does this mean that Google is finding this and thinks that I have only two words (page contents) on the page? Or is this code to grab the page contents from somewhere else in the code? I'm completely lost with this and would appreciate any insight.
Technical SEO | | Photowife1 -
Duplicate Page Content Issue
Hello, I recently solved www / no www duplicate issue for my website, but now I am in trouble with duplicate content again. This time something that I cannot understand happens: In Crawl Issues Report, I received Duplicate Page Content for http://yourappliancerepairla.com (DA 19) http://yourappliancerepairla.com/index.html (DA 1) Could you please help me figure out what is happenning here? By default, index.html is being loaded, but this is the only index.html I have in the folder. And it looks like the crawler sees two different pages with different DA... What should I do to handle this issue?
Technical SEO | | kirupa0 -
Thousands of 404-pages, duplicate content pages, temporary redirect
Hi, i take over the SEO of a quite large e-commerce-site. After checking crawl issues, there seems to be +3000 4xx client errors, +3000 duplicate content issues and +35000 temporary redirects. I'm quite desperate regarding these results. What would be the most effective way to handle that. It's a magento shop. I'm grateful for any kind of help! Thx,
Technical SEO | | posthumus
boris0 -
Is the Authority of Individual Pages Diluted When You Add New Pages?
I was wondering if the authority of individual pages is diluted when you add new pages (in Google's view). Suppose your site had 100 pages and you added 100 new pages (without getting any new links). Would the average authority of the original pages significantly decrease and result in a drop in search traffic to the original pages? Do you worry that adding more pages will hurt pages that were previously published?
Technical SEO | | Charlessipe0 -
Can iFrames count as duplicate content on either page?
Hi All Basically what we are wanting to do is insert an iframe with some text on onto a lot of different pages on one website. Does google crawl the content that is in an iFrame? Thanks
Technical SEO | | cttgroup0 -
Would Google Call These Pages Duplicate Content?
Our Web store, http://www.audiobooksonline.com/index.html, has struggled with duplicate content issues for some time. One aspect of duplicate content is a page like this: http://www.audiobooksonline.com/out-of-publication-audio-books-book-audiobook-audiobooks.html. When an audio book title goes out-of-publication we keep the page at our store and display a http://www.audiobooksonline.com/out-of-publication-audio-books-book-audiobook-audiobooks.html whenever a visitor attempts to visit a specific title that is OOP. There are several thousand OOP pages. Would Google consider these OOP pages duplicate content?
Technical SEO | | lbohen0 -
Does page size and relative content position affect SEO?
Good morning, Each product page of our e-commerce site consists of a fairly lengthy header and footer. The former of which contains links to ~60 product categories, the logo, etc, while the latter contains information such as the latest posts from our blog, links to support, etc. The main "content" of the page is of course product related information, which also happens to contain a bit of templated data such as links which when clicked open respective sliders containing information regarding our return and shipping policies. The question: We wonder whether the relative "size" of the page has anything to do with SEO results. As an example, suppose the page header consists of 20% of the total page size, the important page-specific content consumes 60%, and the footer consumes the final 20%. Is this relevant? Or to rephrase the question: Should we be concerned about keeping our headers and footers as small as possible? Thanks!
Technical SEO | | FondriestEnv0