What is the Googlebot?
The Googlebot is a webCrawling-software search bot (also known as a spider or Webcrawler), which collects the website information used to provide Google search engine results pages (SERP) can be used.
The bot collects documents from the web to build Google's search index. By constantly collecting documents, the software discovers new pages and updates existing pages. Googlebot Uses a distributed design that spans many computers so it can grow with the web.
How does the Googlebot work?
The Webcrawler uses algorithms to determine which pages are to be searched, at which frequency the pages are to be searched and from how many pages information is to be retrieved. The Googlebot starts with a list generated from previous sessions. This list is then supplemented by the sitemaps of the Webmaster supplements. The software scans all linked elements on the websites it searches and notices new pages, page updates and dead links. The information collected is used to optimize the Index from Google on the web.
How can I influence the Google crawler?
The bot creates a Index within the restrictions defined by webmasters in their robots.txt files. If a Webmaster For example, if you want to hide pages from Google searches, the Google bot can block them in a robots.txt file in the top folder of the website.
To prevent Googlebot links on a particular page of a website, he can follow the nofollow-meta tag; to prevent the bot from following individual links, the Webmaster rel="nofollow" to the links themselves.
The Webmaster of a website can recognize every few seconds the visits of computers to google.com that use the user agent Googlebot display. In general, Google tries to index as much of a website as possible without overloading the website's bandwidth. When a Webmaster states that Googlebot consumes too much bandwidth, it can be switched on the Homepage the Search Console from Google set a rate that remains valid for 90 days.
« Back to Glossary Index