Back to Search Start Over

Prior latent distribution comparison for the RNN Variational Autoencoder in low-resource language modeling.

Authors :
Kostiuk, Yevhen
Lukashchuk, Mykola
Gelbukh, Alexander
Sidorov, Grigori
Pinto, David
Beltrán, Beatriz
Singh, Vivek
Source :
Journal of Intelligent & Fuzzy Systems. 2022, Vol. 42 Issue 5, p4541-4549. 9p.
Publication Year :
2022

Abstract

Probabilistic Bayesian methods are widely used in the machine learning domain. Variational Autoencoder (VAE) is a common architecture for solving the Language Modeling task in a self-supervised way. VAE consists of a concept of latent variables inside the model. Latent variables are described as a random variable that is fit by the data. Up to now, in the majority of cases, latent variables are considered normally distributed. The normal distribution is a well-known distribution that can be easily included in any pipeline. Moreover, the normal distribution is a good choice when the Central Limit Theorem (CLT) holds. It makes it effective when one is working with i.i.d. (independent and identically distributed) random variables. However, the conditions of CLT in Natural Language Processing are not easy to check. So, the choice of distribution family is unclear in the domain. This paper studies the priors selection impact of continuous distributions in the Low-Resource Language Modeling task with VAE. The experiment shows that there is a statistical difference between the different priors in the encoder-decoder architecture. We showed that family distribution hyperparameter is important in the Low-Resource Language Modeling task and should be considered for the model training. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
10641246
Volume :
42
Issue :
5
Database :
Academic Search Index
Journal :
Journal of Intelligent & Fuzzy Systems
Publication Type :
Academic Journal
Accession number :
156139436
Full Text :
https://doi.org/10.3233/JIFS-219243