Sorry, I don't understand your search. ×
Back to Search Start Over

On Alignment in Deep Linear Neural Networks

Authors :
Radhakrishnan, Adityanarayanan
Nichani, Eshaan
Bernstein, Daniel
Uhler, Caroline
Publication Year :
2020

Abstract

We study the properties of alignment, a form of implicit regularization, in linear neural networks under gradient descent. We define alignment for fully connected networks with multidimensional outputs and show that it is a natural extension of alignment in networks with 1-dimensional outputs as defined by Ji and Telgarsky, 2018. While in fully connected networks, there always exists a global minimum corresponding to an aligned solution, we analyze alignment as it relates to the training process. Namely, we characterize when alignment is an invariant of training under gradient descent by providing necessary and sufficient conditions for this invariant to hold. In such settings, the dynamics of gradient descent simplify, thereby allowing us to provide an explicit learning rate under which the network converges linearly to a global minimum. We then analyze networks with layer constraints such as convolutional networks. In this setting, we prove that gradient descent is equivalent to projected gradient descent, and that alignment is impossible with sufficiently large datasets.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2003.06340
Document Type :
Working Paper