Why this page doesn't get indexed?
-
Hi, I've just taken over development and SEO for a site and we're having difficulty getting some key pages indexed on our site. They are two clicks away from the homepage, but still not getting indexed. They are recently created pages, with unique content on. The architecture looks like this:Homepage >> Car page >> Engine specific pageWhenever we add a new car, we link to its 'Car page' and it gets indexed very quickly. However the 'Engine pages' for that car don't get indexed, even after a couple of weeks. An example of one of these index pages are - http://www.carbuzz.co.uk/car-reviews/Volkswagen/Beetle-New/2.0-TSISo, things we've checked - 1. Yes, it's not blocked by robots.txt2. Yes, it's in the sitemap (http://www.carbuzz.co.uk/sitemap.xml)3. Yes, it's viewable to search spiders (e.g. the link is present in the html source)This page doesn't have a huge amount of unique content. We're a review aggregator, but it still does have some. Any suggestions as to why it isn't indexed?Thanks, David
-
Hi David,
Apologies, in my haste I didnt take in it was the engine page. I hole heartedly agree with the other comments made here. I would also add that, although not the be all and end all, you may find that linking out to to other sites may help the page get indexed as well. If there are any additional, authoritative resources (that are not competition) it may be worth linking to a few of these. Adds value to the user as well.
-
A factor could be another page was not indexed on the previous version of the site, which is presently indexed. On this site, the "other" page was indexed first, leading to this page not being indexed.
I noticed your site has other 2.0-TSI pages, and they were VW pages as well.
-
Interesting in the previous (completely version of the site) that page used to get crawled fine, but it had more content (but exactly the same as the parent one).
I wonder whether I should accept some duplication, just bite the bullet and write lots of original content, maybe find a way to explain google (using microformats) that it's a page that aggregates reviews.
I still need to figure out whether is it more worthwhile focusing on long tail keywords that the engine page would allow (E.g. Audi A6 1.4 Bluemotion TDI reviews) or just merge it together with the parent one for a much more content rich (Audi A6 Reviews).
-
Completely agree with Ryan, as an in-house SEO of a company with hundreds of thousands of pages I can guarantee you that not every page you have will be indexed via your sitemaps. The best thing to do is just use the tips that Ryan stated above. Link earning or content building, content building being the easier of the two.
-
Google has specifically stated they do not guarantee they will index all pages of a website. For Google to index a page, they have to decide the page offers value.
When I examine the page I note the following:
-
the page's comments are all snippets from other pages which have the full comments. There is no unique content for the comments.
-
the page's content is a total of 6 sentences, several of which are common to other pages on your site or are otherwise generic such as "Use the filter above to see reviews for the other engines."
-
the page has no backlinks to it, and the linking page has no links either. There are no apparent off page factors to indicate this page is important.
If you were to earn links to the page, my bet is it would be indexed. If you were to add a decent amount of quality content on the page, it would probably be indexed as well.
-
-
That's not the problem. The car in question got crawled correctly. It's the engine below it that didn't get crawled.
Anyway the car-chooser page is not the only way to get to cars.
-
Hi David,
When you first load the "Car Page" http://www.carbuzz.co.uk/car-chooser only the first few cars show. THe rest are only accessible when you scroll down the page which then activates some AJAX (i think) which loads more cars i the listing.
I suspect this has something to do with is. The bots cant access the "more" listings.
You could use an HTML sitemap on the site to help them get indexed but it would be better if the pages were linked to by default from the listings page.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How to get google to forget my old but still working page and list my new fully optimized page for a keyword?
Hi There! (i am beginner in seo) I have dynamic and static pages on our site. I created a static page for a specific keyword. Fully optimized it, (h1, alt, metas, etc.....maybe too optimized). My problem is that this page is alive for weeks, checked it in GWT and it is in robots.txt, google sees it, and indexed it. BUT whenewer i do a search for that keyword, we still appear with the dynamically created link in the google listings. How could i "redirect" google, if sy make a search for that keyword than shows our optimized page? Is there a tool for that? I cant delete the dynamic page... Any ideas? Thx Andrew
Technical SEO | | Neckermann0 -
Will a Robots.txt 'disallow' of a directory, keep Google from seeing 301 redirects for pages/files within the directory?
Hi- I have a client that had thousands of dynamic php pages indexed by Google that shouldn't have been. He has since blocked these php pages via robots.txt disallow. Unfortunately, many of those php pages were linked to by high quality sites mulitiple times (instead of the static urls) before he put up the php 'disallow'. If we create 301 redirects for some of these php URLs that area still showing high value backlinks and send them to the correct static URLs, will Google even see these 301 redirects and pass link value to the proper static URLs? Or will the robots.txt keep Google away and we lose all these high quality backlinks? I guess the same question applies if we use the canonical tag instead of the 301. Will the robots.txt keep Google from seeing the canonical tags on the php pages? Thanks very much, V
Technical SEO | | Voodak0 -
New pages need to be crawled & indexed
Hi there, When you add pages to a site, do you need to re-generate an XML site map and re-submit to Google/Bing? I see the option in Google Webmaster Tools under the "fetch as Google tool" to submit individual pages for indexing, which I am doing right now. Thanks,
Technical SEO | | SSFCU
Sarah0 -
How to stop google from indexing specific sections of a page?
I'm currently trying to find a way to stop googlebot from indexing specific areas of a page, long ago Yahoo search created this tag class=”robots-nocontent” and I'm trying to see if there is a similar manner for google or if they have adopted the same tag? Any help would be much appreciated.
Technical SEO | | Iamfaramon0 -
Why wont google Index this page?
A week ago i accidentally changed this page settings in my CMS to "disable & dont index" as i was going to replace this page with another, but this didnt happen, but i forgot to switch the settings back! http://www.over50choices.co.uk/funeral-planning/funeral-plans Anyhow in an effort to get it back up quickly i submitted in GWTs but its still not indexed. When i use several SEO on page checking tools it has the Meta Title data as "Form" and not the correct title. Any ideas please? Yours frustrated Ash
Technical SEO | | AshShep10 -
What coding works for SEO and what coding doesn't?
Hi: I recently learned about inline styles and that Google started penalizing sites for that in October. Then I was told that Wix and Flash don't work (or work well) either for SEO as the engines won't crawl them (I think). Does anyone know of a blog that goes over everything that doesn't work so that I could recognize it when I look at someone's code. Anyone know of such a resource? Cheers, Wes
Technical SEO | | wrconard0 -
Error msg 'Duplicate Page Content', how to fix?
Hey guys, I'm new to SEO and have the following error msg 'Duplicate Page Content'. Of course I know what it means, but my question is how do you delete the old pages that has duplicate content? I use to run my website through Joomla! but have since moved to Shopify. I see that the duplicated site content is still from the old Joomla! site and I would like to learn how to delete this content (or best practice in this situation). Any advice would be very helpful! Cheers, Peter
Technical SEO | | pjuszczynski0 -
Mask links with JS that point to noindex'ed paged
Hi, in an effort to prepare our page for the Panda we dramatically reduced the number of pages that can be indexed (from 100k down to 4k). All the remaining pages are being equipped with unique and valuable content. We still have the other pages around, since they represent searches with filter combination which we deem are less interesting to the majority of users (hence they are not indexed). So I am wondering if we should mask links to these non-indexed pages with JS, such that Link-Juice doesn't get lost to those. Currently the targeted pages are non-index via "noindex, follow" - we might de-index them with robots.txt though, if the "site:" query doesn't show improvements. Thanks, Sebastian
Technical SEO | | derderko0