Analyzing the web crawler as a feed forward engine for an efficient solution to the search problem in the minimum amount of time through a distributed framework

M. Atif Qureshi, Arjumand Younus, Francisco Rojas

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

A web crawler forms the backbone of a search engine and this backbone needs a careful re-assessment that could enhance the efficiency of search engines. This paper conducts such a re-assessment from the perspective of systems and this is achieved through implementation and analysis of a web crawler "VisionerBOT" as a feed forward engine for search engines using the MapReduce distributed programming model. Our crawler implementations revisit the classical OS debate of threads vs. events, with a significant contribution from our work which concludes that events is the ideal way forward for web crawlers. Furthermore, in implementing the feed forward mechanisms within the web crawler, we came up with some important design considerations for the operating system research community which can lead to a whole new class of operating systems.

Original languageEnglish
Title of host publication2010 International Conference on Information Science and Applications, ICISA 2010
DOIs
Publication statusPublished - 2010
Externally publishedYes
Event2010 International Conference in Information Science and Applications, ICISA 2010 - Seoul, Korea, Republic of
Duration: 21 Apr 201023 Apr 2010

Publication series

Name2010 International Conference on Information Science and Applications, ICISA 2010

Conference

Conference2010 International Conference in Information Science and Applications, ICISA 2010
Country/TerritoryKorea, Republic of
CitySeoul
Period21/04/1023/04/10

Keywords

  • Events
  • Feed forward
  • MapReduce
  • Threads
  • Web crawler

Fingerprint

Dive into the research topics of 'Analyzing the web crawler as a feed forward engine for an efficient solution to the search problem in the minimum amount of time through a distributed framework'. Together they form a unique fingerprint.

Cite this