2014 ReducingtheSamplingComplexityof

From GM-RKB
Jump to navigation Jump to search

Subject Headings:

Notes

Cited By

Quotes

Author Keywords

Abstract

Inference in topic models typically involves a sampling step to associate latent variables with observations. Unfortunately the generative model loses sparsity as the amount of data increases, requiring O (k) operations per word for k topics. In this paper we propose an algorithm which scales linearly with the number of actually instantiated topics kd in the document. For large document collections and in structured hierarchical models kd ll k. This yields an order of magnitude speedup. Our method applies to a wide variety of statistical models such as PDP [16, 4] and HDP [19].

At its core is the idea that dense, slowly changing distributions can be approximated efficiently by the combination of a Metropolis-Hastings step, use of sparsity, and amortized constant time sampling via Walker's alias method.

References

;

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2014 ReducingtheSamplingComplexityofAlexander J. Smola
Amr Ahmed
Aaron Q. Li
Sujith Ravi
Reducing the Sampling Complexity of Topic Models10.1145/2623330.26237562014