Your site has a scheduled crawl frequency... maybe once per month, once per week, once per day, once per hour... depending in large part on how often Google finds that you generate new content. This is when they typically deep crawl and request the most pages on your site.
Google also crawls your site in between these scheduled crawl times because it's crawling other site. When those sites link to you Google follows those links to find out if the page on your site they are linking to still exists, or if it's 404ing, or if it's been 301 redirected. When the crawler visits because of one of these inbound links to your sites, they often will crawl a small number of pages on your site.
Sites with high PR typically have lots of inbound links which is why those sites get crawled frequently.
I was SEO for a PR7 site w/ 4.7 million inbound links according to Yahoo! Site Explorer and Google was crawling that site 24 hrs per day.
When best bodies allocution about Internet chase engines, they absolutely beggarly World Wide Web chase engines. Before the Web became the best arresting allotment of the Internet, there were already chase engines in abode to advice bodies acquisition advice on the Net. Programs with names like "gopher" and "Archie" kept indexes of files stored on servers affiliated to the Internet, and badly bargain the bulk of time appropriate to acquisition programs and documents. In the backward 1980s, accepting austere amount from the Internet meant alive how to use gopher, Archie, Veronica and the rest.
Googlebot is Google’s web ample robot, which finds and retrieves pages on the web and easily them off to the Google indexer. It’s accessible to brainstorm Googlebot as a little spider scurrying beyond the strands of cyberspace, but in absoluteness Googlebot doesn’t bisect the web at all. It functions abundant like your web browser, by sending a appeal to a web server for a web page, downloading the absolute page, again handing it off to Google’s indexer.
Googlebot consists of abounding computers requesting and attractive pages abundant added bound than you can with your web browser. In fact, Googlebot can appeal bags of altered pages simultaneously. To abstain cutting web servers, or bottleneck out requests from animal users, Googlebot advisedly makes requests of anniversary alone web server added boring than it’s able of doing.
Unfortunately, spammers ample out how to actualize automatic bots that bombarded the add URL anatomy with millions of URLs pointing to bartering propaganda. Google rejects those URLs submitted through its Add URL anatomy that it suspects are aggravating to deceive users by employing approach such as including hidden argument or links on a page, capacity a folio with extraneous words, cloaking (aka allurement and switch), application base redirects, creating doorways, domains, or sub-domains with essentially agnate content, sending automatic queries to Google, and bond to bad neighbors. So now the Add URL anatomy additionally has a test: it displays some squiggly belletrist advised to fool automatic “letter-guessers”; it asks you to access the belletrist you see — article like an eye-chart analysis to stop spambots.
When Googlebot fetches a page, it culls all the links actualization on the folio and adds them to a chain for consecutive crawling. Googlebot tends to appointment little spam because best web authors articulation alone to what they accept are high-quality pages. By agriculture links from every folio it encounters, Googlebot can bound body a account of links that can awning ample alcove of the web.
This technique, accepted as abysmal crawling, additionally allows Googlebot to delving abysmal aural alone sites. Because of their massive scale, abysmal crawls can ability about every folio in the web. Because the web is vast, this can booty some time, so some pages may be crawled alone already a month.
Although its action is simple, Googlebot charge be programmed to handle several challenges. First, back Googlebot sends out accompanying requests for bags of pages, the chain of “visit soon” URLs charge be consistently advised and compared with URLs already in Google’s index. Duplicates in the chain charge be alone to anticipate Googlebot from attractive the aforementioned folio again. Googlebot charge actuate how generally to revisit a page. On the one hand, it’s a decay of assets to re-index an banausic page. On the added hand, Google wants to re-index afflicted pages to bear abreast results.
To accumulate the basis current, Google continuously recrawls accepted frequently alteration web pages at a amount almost proportional to how generally the pages change. Such crawls accumulate an basis accepted and are accepted as beginning crawls. Newspaper pages are downloaded daily, pages with banal quotes are downloaded abundant added frequently. Of course, beginning crawls acknowledgment beneath pages than the abysmal crawl. The aggregate of the two types of crawls allows Google to both accomplish able use of its assets and accumulate its basis analytic current.
Googlebot gives the indexer the abounding argument of the pages it finds. These pages are stored in Google’s basis database. This basis is sorted alphabetically by chase term, with anniversary basis admission autumn a account of abstracts in which the appellation appears and the area aural the argument area it occurs. This abstracts anatomy allows accelerated admission to abstracts that accommodate user concern terms.
Google’s Concern Processor
The concern processor has several parts, including the user interface (search box), the “engine” that evaluates queries and matches them to accordant documents, and the after-effects formatter.
Post Source by