Abstract
By representing semantics in latent spaces, Variational autoencoders (VAEs) have been proven powerful in modelling and generating signals such as image and text, even without supervision. However, previous studies suggest that in a learned latent space, some low density regions (aka. holes) exist, which could harm the overall system performance. While existing studies focus on empirically mitigating these latent holes, how they distribute and how they affect different components of a VAE, are still unexplored. In addition, the hole issue in VAEs for language processing is rarely addressed. In our work, by introducing a simple hole-detection algorithm based on the neighbour consistency between VAE’s input, latent, and output semantic spaces, we propose to deeply dive into these topics for the first time. Comprehensive experiments including automatic evaluation and human evaluation imply that large-scale low-density latent holes may not exist in the latent space. In addition, various sentence encoding strategies are explored and the native word embedding is the most suitable strategy for VAEs in language modelling task.
Original language | English |
---|---|
Title of host publication | Proceedings of Machine Learning Research |
Subtitle of host publication | NeurIPS 2020 Preregistration Workshop |
Pages | 343-357 |
Number of pages | 15 |
Volume | 148 |
Publication status | Published - 8 Jul 2021 |
Externally published | Yes |
Event | NeurIPS 2020 Workshop on Pre-registration in Machine Learning - Virtual event Duration: 11 Dec 2020 → 11 Dec 2020 https://preregister.science/neurips2020.html |
Publication series
Name | Proceedings of Machine Learning Research |
---|---|
Publisher | MLResearchPress |
ISSN (Electronic) | 2640-3498 |
Workshop
Workshop | NeurIPS 2020 Workshop on Pre-registration in Machine Learning |
---|---|
Period | 11/12/20 → 11/12/20 |
Internet address |
Keywords
- variational autoencoder
- low-density regions
- latent holes