[Paper Reding] Discovering Discrete Latent Topics with Neural Variational Inference

Ya-Liang Allen Chang
3 min readMay 2, 2019

--

Problem Definition

Develop a fast, accurate and expressive topic model for documents

Introduction

Latent semantic analysis (LSA)

Probabilistic topic models (e.g. PLSA , LDA and HDPs)

  • Robust, scalable, and theoretically sound foundation for document modeling
  • Introduce latent variables for each token to topic assignment

Traditional Dirichlet-Multinomial topic model

  • Inference becomes more complex as topic models have grown more expressive

Deep neural networks

  • Freat potential for learning complicated non-linear distributions for unsupervised models
  • Neural variational inference
  • Continuous/discrete latent variables

Contributions

  • Propose and evaluate a range of topic models parameterised with neural networks and trained with variational inference
  • Introduce three different neural structures for constructing topic distributions: the Gaussian Softmax distribution (GSM), the Gaussian Stick Breaking distribution (GSB), and the Recurrent Stick Breaking process (RSB)
  • Combine the merits of both neural networks and traditional probabilistic topic models
  • Interpretable and explicitly represent the dependencies amongst the random variables

Parameterising Topic Distributions

Probabilistic topic models, such as LDA

  • Use the latent variables θd and zn for the topic proportion of document d, and the topic assignment for the observed word wn, respectively

Proposed

  • Introduce a neural network to parameterise the multinomial topic distribution

The Gaussian Softmax Construction

The Gaussian Stick Breaking Construction

The Recurrent Stick Breaking Construction

Models

Neural Topic Models

Recurrent Neural Topic Models

Topic vs. Document Models

Experiments

Dataset

MXM2 song lyrics, 20NewsGroups3 and Reuters RCV1-v24 news

Evaluation

Document perplexity

Results

Summary

In this paper we have introduced a family of neural topic models using the Gaussian Softmax, Gaussian StickBreaking and Recurrent Stick-Breaking constructions for parameterising the latent multinomial topic distributions of
each document.

Reference

Miao, Yishu, Edward Grefenstette, and Phil Blunsom. “Discovering discrete latent topics with neural variational inference.” Proceedings of the 34th International Conference on Machine Learning-Volume 70. JMLR. org, 2017.

--

--

No responses yet