by A. Gotovos, H. Hassani, A. Krause, S. Jegelka

Abstract:

We consider the problem of inference in discrete probabilistic models, that is, distributions over subsets of a finite ground set. These encompass a range of well-known models in machine learning, such as determinantal point processes and Ising models. Locally-moving Markov chain Monte Carlo algorithms, such as the Gibbs sampler, are commonly used for inference in such models, but their convergence is, at times, prohibitively slow. This is often caused by state-space bottlenecks that greatly hinder the movement of such samplers. We propose a novel sampling strategy that uses a specific mixture of product distributions to propose global moves and, thus, accelerate convergence. Furthermore, we show how to construct such a mixture using semigradient information. We illustrate the effectiveness of combining our sampler with existing ones, both theoretically on an example model, as well as practically on three models learned from real-world data sets.

Reference:

Discrete Sampling using Semigradient-based Product Mixtures A. Gotovos, H. Hassani, A. Krause, S. JegelkaIn Conference on Uncertainty in Artificial Intelligence (UAI), 2018Oral presentation

Bibtex Entry:

@inproceedings{gotovos18discrete, Author = {Alkis Gotovos and Hamed Hassani and Andreas Krause and Stefanie Jegelka}, Booktitle = {Conference on Uncertainty in Artificial Intelligence (UAI)}, Month = {August}, Title = {Discrete Sampling using Semigradient-based Product Mixtures}, Year = {2018}}