by , , ,
Many large-scale machine learning problems (such as clustering, non-parametric learning, kernel machines, etc.) require selecting, out of a massive data set, a manageable yet representative subset. Such problems can often be reduced to maximizing a submodular set function subject to cardinality constraints. Classical approaches require centralized access to the full data set; but for truly large-scale problems, rendering the data centrally is often impractical. In this paper, we consider the problem of submodular function maximization in a distributed fashion. We develop a simple, two-stage protocol GREEDI, that is easily implemented using MapReduce style computations. We theoretically analyze our approach, and show, that under certain natural conditions, performance close to the (impractical) centralized approach can be achieved. In our extensive experiments, we demonstrate the effectiveness of our approach on several applications, including sparse Gaussian process inference and exemplar-based clustering, on tens of millions of data points using Hadoop.
Distributed Submodular Maximization: Identifying Representative Elements in Massive Data B. Mirzasoleiman, A. Karbasi, R. Sarkar, A. KrauseIn Neural Information Processing Systems (NIPS), 2013
Bibtex Entry:
	author = {Baharan Mirzasoleiman and Amin Karbasi and Rik Sarkar and Andreas Krause},
	booktitle = {Neural Information Processing Systems (NIPS)},
	month = {December},
	title = {Distributed Submodular Maximization: Identifying Representative Elements in Massive Data},
	year = {2013}}