Back to Search Start Over

Regularized linear autoencoders recover the principal components, eventually

Authors :
Bao, Xuchan
Lucas, James
Sachdeva, Sushant
Grosse, Roger
Source :
Advances in Neural Information Processing Systems 33 (NeurIPS 2020)
Publication Year :
2020

Abstract

Our understanding of learning input-output relationships with neural nets has improved rapidly in recent years, but little is known about the convergence of the underlying representations, even in the simple case of linear autoencoders (LAEs). We show that when trained with proper regularization, LAEs can directly learn the optimal representation -- ordered, axis-aligned principal components. We analyze two such regularization schemes: non-uniform $\ell_2$ regularization and a deterministic variant of nested dropout [Rippel et al, ICML' 2014]. Though both regularization schemes converge to the optimal representation, we show that this convergence is slow due to ill-conditioning that worsens with increasing latent dimension. We show that the inefficiency of learning the optimal representation is not inevitable -- we present a simple modification to the gradient descent update that greatly speeds up convergence empirically.

Details

Database :
arXiv
Journal :
Advances in Neural Information Processing Systems 33 (NeurIPS 2020)
Publication Type :
Report
Accession number :
edsarx.2007.06731
Document Type :
Working Paper