Topic Embeddings

Contextualized Topic Models

Introduced by Bianchi et al. in Cross-lingual Contextualized Topic Models with Zero-shot Learning

Contextualized Topic Models are based on the Neural-ProdLDA variational autoencoding approach by Srivastava and Sutton (2017).

This approach trains an encoding neural network to map pre-trained contextualized word embeddings (e.g., BERT) to latent representations. Those latent representations are sampled variationally from a Gaussian distribution $N(\mu, \sigma^2)$ and passed to a decoder network that has to reconstruct the document bag-of-word representation.

Source: Cross-lingual Contextualized Topic Models with Zero-shot Learning

Papers


Paper Code Results Date Stars

Categories