1. Improving representation learning through variational autoencoding
- Author
-
Lin, Shuyu, Trigoni, Niki, and Roberts, Stephen
- Subjects
Machine learning ,Deep learning (Machine learning) - Abstract
Representation learning aims to distill useful knowledge from raw data and apply this knowledge to a wide range of applications. This ability to extract information that is useful not only for selected tasks but also generalizes to new settings is a key step towards artificial intelligence. In this thesis, we focus on representations derived through a specific type of generative model, i.e. variational autoencoders (VAEs). VAEs have several desirable properties. Thanks to the use of variational inference and the convenient model assumptions of Gaussian posteriors and a simple prior, VAEs are often easy to train and exhibit fast convergence. The probabilistic modelling formulation allows VAEs to derive a smooth latent representation of the raw data (i.e. semantically similar data samples are likely to be projected to nearby regions in the latent space). VAEs compress the raw data to a much lower dimension latent space. Working with the low-dimensional representations rather than the raw data can significantly reduces costs in memory and computation. With these advantages, VAEs have been widely applied to many applications, including robotics, drug discovery and digital content creation. Despite the widespread application of VAEs, improving the generative modelling of VAEs further remains an active research topic. In this thesis, we focus on two challenges in the VAE training: 1) over-regularized posterior distributions are often encountered in VAEs with Gaussian decoders and simple prior models; 2) the auto-encoding function may cause severe information drift and alter the information in the raw data in successive encodings. We propose solutions to both phenomena. Specifically, we optimize a variance parameter in the Gaussian decoder to balance competing loss terms in the ELBO objective. We adopt a flexible prior model that is implemented as a VAE in the latent space to mitigate the over-regularization effects. To reduce the information drift, we propose to modify the ELBO objective with a consistency loss that penalizes such drift. We show that these proposals can effectively address the challenges identified previously and improve the likelihood score of the VAEs. In addition to the contributions related to improving VAEs, we also demonstrate the power of VAEs' representation learning in two important machine learning applications. Firstly, we show that a VAE's ability to compress complicated, high-dimensional data is key to achieving good performance in anomaly detection. We design a VAE-LSTM anomaly detection system that can accurately identify anomalous effects in a time serious. Secondly, we show that a classifier which incorporates a VAE module can give better calibrated predictions. This is the result of a VAEs' capability of expressing the uncertainty between similar data samples in the spread of the posterior distribution and of identifying out-of-distribution samples.
- Published
- 2022