Getting pages indexed by Google
-
Hi SEOMoz,
I relaunched a site back in February of this year (www.uniquip.com) with about 1 million URL's.
Right now I'm seeing that Google is not going past 110k indexed URL's (based on sitemaps).
Do you have any tips on what I can do to make the site more likeable by Google and get more indexed URL's?
All the the part pages can be browsed to by going to: http://www.uniquip.com/product-line-card/suppliers/sw-a/p-1
I've tried to make the content as unique as possible by adding random testimonials and random "related part numbers" see here: http://www.uniquip.com/id/246172/electronic-components/infineon/microcontrollers-mcu/sabc161pilfca
Do I need to wait more time and be more patient with Google? It just seems like I'm only getting a few thousand URL's per day at the most.
Would it help me if I implemented a breadcrumb on all part pages?
Thanks,
-Carlos
-
Carlos
Actually, it is in the index but if you do: site:www.uniquip.com as a search in Google it is not the first result, which it should be. If you do "site:www.uniquip.com" (with the quotes) as a search in Google, it is still only the second result, it should be the first, which is a sign that something is weird, which would require some more investigation.
-Dan
-
Thanks for both tips Dan, I will look into them althought number 2 may be a bit tough.... since there are hundreds of suppliers/categories (bit tough to put in a menu).
as for the page not being indexed I'm showing that it is. http://webcache.googleusercontent.com/search?sourceid=navclient-ff&ie=UTF-8&q=cache%3Ahttp%3A%2F%2Fwww.uniquip.com%2F
is this not the case for you?
-
Thanks for the feedback Brent.
I already my sitemaps setup in groups of 30k with a master sitemap file. http://www.uniquip.com/sm/smindex.xml
I will talk to the developer so we can look into regrouping them and perhaps pickup an indexing pattern.
-
Hi Carlos
It appears there are two major issues with your site not getting indexed.
1. Off-Site - Go to opensiteexplorer and type in your URL. You will see there are many suspicious looking links pointing back to your domain. No matter what you do on-site, if you have really low quality backlinks pointing to your domain (and no good ones to balance things out), nothing you do on-site will help. You'll want to do whatever you can to clean these up (get them removed). You'll also want to work on getting some GOOD backlinks in conjunction with cleaning out the bad ones. There are many resources here on SEOmoz on backlinking strategies.
2. On-site - Your product pages are 4 or more clicks away from the homepage. You really want this to be 3 or less (ideally 1 if possible for some products and categories).
I also noticed that your homepage does not appear to be indexed in Google.
We could go WAY more in depth with this, but in general those seem to be the two major issues.
-Dan
-
Carlos,
There is a great thread you should read about this topic: http://www.seomoz.org/q/can-you-push-too-many-urls-via-sitemaps
As for me, I would recommend creating multiple XML sitemaps where you can track which urls are bing indexed and to also control the submission quantity. You don't want to publish a ton of urls at one time. For the urls that aren't in Google's index, I would no index them for now and slowly submit a small group on a monthly basis, along with different XML sitemaps for tracking.
You have to think as if you were Google, is the content you're submitting valuable and new information? There is no way that 1 million new pages is so valuable or new that isn't already out there. A nice steady submission stream will give you the best results in the long run.
I hope this helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What to do about pages I have deleted?
I have been working through the dead links on my page and recreating the page with new content for those pages that it still makes sense to have on the site. But I have a few that were just changes of the title, spelling mistakes or other ways of saying the same thing In other words I created a page called "areas of the UK we cover" but decided to change it to "areas covered" However, I must have created links to this page and now it is a dead link with a page authority of 19 I think it would be spammy to have two pages, one called "areas covered" and the other called "areas of the UK we cover. It's not a disallow in Robots.txt because the page does not exist Please note I do not have access to the header to add code for a 301 redirect. I'm still using webs.com but not for new sites. I also have a page called singing telegrams london, that I changed from singagrams london. These are two words for the same thing but they are two very different keywords would it be ok to recreate this page and create content for singagrams london. Help is much appreciated
On-Page Optimization | | singingtelegramsuk0 -
No-index all the posts of a category
Hi everyone! I would like no-indexing all the posts of a specific category of my wordpress site. The problem is that the structure of my URL is composed without /category/: www.site-name.ext/date/post-name/
On-Page Optimization | | salvyy
so without /category-name/ Is possibile to disallow the indexing of all the posts of the category via robots.txt? Using Yoast Plugin I can put the no-index for each post, but I would like to put the no-index (or disallow/) a time for all the post of the category. Thanks in advance for your help and sorry for my english. Mike0 -
Locating Duplicate Pages
Hi, Our website consists of approximately 15,000 pages however according to our Google Webmaster Tools account Google has around 26,000 pages for us in their index. I have run through half a dozen sitemap generators and they all only discover the 15,000 pages that we know about. I have also thoroughly gone through the site to attempt to find any sections where we might be inadvertently generating duplicate pages without success. It has been over six months since we did any structural changes (at which point we did 301's to the new locations) and so I'd like to think that the majority of these old pages have been removed from the Google Index. Additionally, the number of pages in the index doesn't appear to be going down by any discernable factor week on week. I'm certain it's nothing to worry about however for my own peace of mind I'd like to just confirm that the additional 11,000 pages are just old results that will eventually disappear from the index and that we're not generating any duplicate content. Unfortunately there doesn't appear to be a way to download a list of the 26,000 pages that Google has indexed so that I can compare it against our sitemap. Obviously I know about site:domain.com however this only returned the first 1,000 results which all checkout fine. I was wondering if anybody knew of any methods or tools that we could use to attempt to identify these 11,000 extra pages in the Google index so we can confirm that they're just old pages which haven’t fallen out of the index yet and that they’re not going to be causing us a problem? Thanks guys!
On-Page Optimization | | ChrisHolgate0 -
Why are some of page indexed and others not
I have created a site structure like this: domain/for-sale/brand domain/for-sale/brand-model domain/for-sale/brand-model/pg1 domain/for-sale/brand-model/pg2 domain/for-sale/brand-model/pg3 etc.... I cannot understand why the domain/for-sale/brand-model does not seem to be indexed, yet the domain/for-sale/brand-model/pg6 is? This is a new site, but I cannot understand why this URL would be indexed without the others... Any ideas? My home pages has links to the domain/for-sale/brand, this page has links to domain/for-sale/brand-model1, domain/for-sale/brand-model2 etc, each of these pages have links to domain/for-sale/brand-model/pg1, domain/for-sale/brand-model/pg2 etc...
On-Page Optimization | | MirandaP0 -
Product page optimalisation
Throughout the years our website kept on growing this has led to product pages that have so much sub-pages that nobody is able to really get a good idea of the product. We are working on a new website where the visitor is central. Together with a usability partner we have down sized the preferred data to fit on one page with a tabular system with a maximum of 4 taps. My question will this affect our find ability if we go from 10 to 15 sub-pages to one main page
On-Page Optimization | | TiasNimbas0 -
Creating New Pages Versus Improving Existing Pages
What are some things to consider or things to evaluate when deciding whether you should focus resources on creating new pages (to cover more related topics) versus improving existing pages (adding more useful information, etc.)?
On-Page Optimization | | SparkplugDigital0 -
Got loads of pages, but none indexing?
I have a WordPress site with loads of pages on a url like this http://mysite.com.au However, Google has indexed http://www.mysite.com.au and as a result only indexing 2 pages. How do I fix this? Many thanks Dan
On-Page Optimization | | Pokodot0