TitleThe HaLoop approach to large-scale iterative data analysis
Publication TypeJournal Article
Year of Publication2012
AuthorsBu Y, Howe B, Balazinska M, Ernst MD
JournalThe VLDB Journal
Volume21
Pagination169–190
AbstractThe growing demand for large-scale data mining and data analysis applications has led both industry and academia to design new types of highly scalable data-intensive computing platforms. MapReduce has enjoyed particular success. However, MapReduce lacks built-in support for iterative programs, which arise naturally in many applications including data mining, web ranking, graph analysis, and model fitting. This paper (This is an extended version of the VLDB 2010 paper ``HaLoop: Efficient Iterative Data Processing on Large Clusters'' PVLDB 3(1):285–296, 2010.) presents HaLoop, a modified version of the Hadoop MapReduce framework, that is designed to serve these applications. HaLoop allows iterative applications to be assembled from existing Hadoop programs without modification, and significantly improves their efficiency by providing inter-iteration caching mechanisms and a loop-aware scheduler to exploit these caches. HaLoop retains the fault-tolerance properties of MapReduce through automatic cache recovery and task re-execution. We evaluated HaLoop on a variety of real applications and real datasets. Compared with Hadoop, on as much data between mappers and reducers in the applications that we tested.
Downloadshttps://homes.cs.washington.edu/~mernst/pubs/haloop-vldb2010-slides.pdf VLDB 2010 slides (PDF) https://homes.cs.washington.edu/~mernst/pubs/haloop-vldb2010-slides.ppt VLDB 2010 slides (PowerPoint) https://code.google.com/archive/p/haloop/ HaLoop implementation https://homes.cs.washington.edu/~mernst/pubs/haloop-vldb2012.pdf PDF
Citation KeyBuHBErnst2012