Skip to main content

How to increase pages indexed

There is a 10 ways to increase pages indexed. They are..

1) PageRank
2) Links
3) Sitemap
4) Speed
5) Google's crawl caching proxy
6) Verify
7) Content
9) Staggered launch
10)Size matters.

PageRank

It depends a lot on PageRank. The higher your PageRank the more pages that will be indexed. PageRank isn't a blanket number for all your pages. Each page has its own PageRank. A high PageRank gives the Googlebot more of a reason to return. Matt Cutts confirms, too, that a higher PageRank means a deeper crawl.

Links

Give the Googlebot something to follow. Links (especially deep links) from a high PageRank site are golden as the trust is already established.

Internal links can help, too. Link to important pages from your homepage. On content pages link to relevant content on other pages.

Sitemap

A lot of buzz around this one. Some report that a clear, well-structured Sitemap helped get all of their pages indexed. Google's Webmaster guidelines recommends submitting a Sitemap file

That page has other advice for improving crawlability, like fixing violations and validating robots.txt.

Some recommend having a Sitemap for every category or section of a site.

Speed

A recent O'Reilly report indicated that page load time and the ease with which the Googlebot can crawl a page may affect how many pages are indexed. The logic is that the faster the Googlebot can crawl, the greater number of pages that can be indexed.

This could involve simplifying the structures and/or navigation of the site. The spiders have difficulty with Flash and Ajax. A text version should be added in those instances.

Google's crawl caching proxy

Matt Cutts provides diagrams of how Google's crawl caching proxy at his blog. This was part of the Big Daddy update to make the engine faster. Any one of three indexes may crawl a site and send the information to a remote server, which is accessed by the remaining indexes (like the blog index or the AdSense index) instead of the bots for those indexes physically visiting your site. They will all use the mirror instead.

Verify

Verify the site with Google using the Webmaster tools.

Content

Make sure content is original. If a verbatim copy of another page, the Googlebot may skip it. Update frequently. This will keep the content fresh. Pages with an older timestamp might be viewed as static, outdated, or already indexed.

Staggered launch

Launching a huge number of pages at once could send off spam signals. In one forum, it is suggested that a webmaster launch a maximum of 5,000 pages per week.

Size matters

If you want tens of millions of pages indexed, your site will probably have to be on an Amazon.com or Microsoft.com level.

Know how your site is found, and tell Google

Find the top queries that lead to your site and remember that anchor text helps in links. Use Google's tools to see which of your pages are indexed, and if there are violations of some kind. Specify your preferred domain so Google knows what to index.

Comments

interactive said…
You itemize it well. specially the ways to increase index.....
Lekhni said…
Thanks :) I find your tips quite helpful..

Popular posts from this blog

Top 10 Things to do Before the Spiders come

1: Check the Title Tags. In our opinion, creating great title tags can be one of the most valuable ways to help increase your SEO. Make sure you have relevant, keyword-rich title tags for each page you create. This will make it easy for web browsers to search and display your website. 2: If you have a flash introduction, be sure there is a link that allows you to pass it. Many webpages have a fancy splash page, but no way to navigate around it. Google cannot read into a flash page, so be sure to include a text link to your website’s second index. Overall this type of intro can add to your site's visual appeal, but don't let it ruin your opportunity to get indexed quickly. 3: Don't forget to check the META Tags. When there is a lot of navigation code that wouldn’t make sense to a human searcher, Google relies on these tags to describe a site. Be sure to set up some valid keywords and a description- just in case. 4: Make sure all your links are in working order. Check and dou...

Google Filters

I have been doing SEO for some time now and I have been witness to many a strange occurrence regarding serps. Most of these weird occurrence I would have to say are directly attributed to a Google Filter or Google penalty. So I have been inspired by a post over at webmasterworld and as far as I know there is not a current list out online that list’s all of the potential Google penalties so I have decided to put together an arbitrary list of potential Google Penalties. Please note that there is no proof i.e. press release from Google stating these exist but rather these are ideas, theories and assumptions from SEO’s experiences. Google Sandbox: The Sandbox Filter is usually applied to brand new websites but has been seen to be applied to domains that have been around for a while. Since most websites do not make it past a year Google implemented a filter that will prevent a new site from getting decent rankings for competitive keyword terms. Usually brand new sites can still rank for ...