Authors
Silvio Lattanzi, Benjamin Moseley, Siddharth Suri, Sergei Vassilvitskii
Publication date
2011/6/4
Book
Proceedings of the twenty-third annual ACM symposium on Parallelism in algorithms and architectures
Pages
85-94
Description
The MapReduce framework is currently the de facto standard used throughout both industry and academia for petabyte scale data analysis. As the input to a typical MapReduce computation is large, one of the key requirements of the framework is that the input cannot be stored on a single machine and must be processed in parallel. In this paper we describe a general algorithmic design technique in the MapReduce framework called filtering. The main idea behind filtering is to reduce the size of the input in a distributed fashion so that the resulting, much smaller, problem instance can be solved on a single machine. Using this approach we give new algorithms in the MapReduce framework for a variety of fundamental graph problems for sufficiently dense graphs. Specifically, we present algorithms for minimum spanning trees, maximal matchings, approximate weighted matchings, approximate vertex and edge …
Total citations
20112012201320142015201620172018201920202021202220232024315253530212433332722211212
Scholar articles
S Lattanzi, B Moseley, S Suri, S Vassilvitskii - Proceedings of the twenty-third annual ACM …, 2011