On the Low-density Latent Regions of VAE-based Language Models

Ruizhe Li* (Corresponding Author), Xutan Peng* (Corresponding Author), Chenghua Lin, Wenge Rong, Zhigang Chen

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

Abstract

By representing semantics in latent spaces, Variational autoencoders (VAEs) have been proven powerful in modelling and generating signals such as image and text, even without supervision. However, previous studies suggest that in a learned latent space, some low density regions (aka. holes) exist, which could harm the overall system performance. While existing studies focus on empirically mitigating these latent holes, how they distribute and how they affect different components of a VAE, are still unexplored. In addition, the hole issue in VAEs for language processing is rarely addressed. In our work, by introducing a simple hole-detection algorithm based on the neighbour consistency between VAE’s input, latent, and output semantic spaces, we propose to deeply dive into these topics for the first time. Comprehensive experiments including automatic evaluation and human evaluation imply that large-scale low-density latent holes may not exist in the latent space. In addition, various sentence encoding strategies are explored and the native word embedding is the most suitable strategy for VAEs in language modelling task.
Original languageEnglish
Title of host publicationProceedings of Machine Learning Research
Subtitle of host publicationNeurIPS 2020 Preregistration Workshop
Pages343-357
Number of pages15
Volume148
Publication statusPublished - 2021

Bibliographical note

Acknowledgement
This work is supported by the award made by the UK Engineering and Physical Sciences
Research Council (Grant number: EP/P011829/1) and Ningbo Natural Science Foundation
(202003N4320, 202003N4321). We would like to thank all the anonymous reviewers for their
insightful and helpful comments.

Fingerprint

Dive into the research topics of 'On the Low-density Latent Regions of VAE-based Language Models'. Together they form a unique fingerprint.

Cite this