Monday, December 27, 2010

How Google Works (Crawling, Indexing)


Crawling

Crawling is the process by which Googlebot discovers new and updated pages to be added to the Google index.

We use a huge set of computers to fetch (or "crawl") billions of pages on the web. The program that does the fetching is called Googlebot (also known as a robot, bot, or spider). Googlebot uses an algorithmic process: computer programs determine which sites to crawl, how often, and how many pages to fetch from each site.

Google's crawl process begins with a list of web page URLs, generated from previous crawl processes, and augmented with Sitemap data provided by webmasters. As Googlebot visits each of these websites it detects links on each page and adds them to its list of pages to crawl. New sites, changes to existing sites, and dead links are noted and used to update the Google index.

Google doesn't accept payment to crawl a site more frequently, and we keep the search side of our business separate from our revenue-generating AdWords service.

Indexing
Googlebot processes each of the pages it crawls in order to compile a massive index of all the words it sees and their location on each page. In addition, we process information included in key content tags and attributes, such as Title tags and ALT attributes. Googlebot can process many, but not all, content types. For example, we cannot process the content of some rich media files or dynamic pages.

Wednesday, December 22, 2010

Google Holds on SEO

Google holds over 60% of the total search market.[38] Its algorithm is naturally also unique, so ranking on Google carries its own unique considerations. Although there are over 200 criteria Google uses to rank sites, they can be categorized into two main sections: on-site and off-site factors:

Google values sites that deliver quality content, relevance, easy navigation and load and an overall user-friendliness to the site’s visitors (on-site). However, a site’s popularity is heavily weighted when Google ranks sites (off-site).[39] Thus Google was originally designed to rank sites mostly based on the number of inbound links they were receiving from other sites. In other words, the more site A was used as a “reference” the higher it would rank. Anchored text links used to link to site A are also very important as well as the popularity and the relevance of the site that is referencing site A.




http://www.heliumx.com

Internet Marketing Services Pakistan

Heliumx SEO Services Pakistan is 100% Ethical (White Hat) SEO Company that follows Search Engine Guidelines and Instructions, we don’t go for any SPAM (Black hat SEO Technique) for us and for our Clients.
We’ve Proven History of Successful Search Engine Optimization (SEO) Techniques since 2003.




Http://www.heliumx.com