1. On the expected behaviour of noise regularised deep neural networks as Gaussian processes.
- Author
-
Pretorius, Arnu, Kamper, Herman, and Kroon, Steve
- Subjects
- *
GAUSSIAN processes , *NOISE , *SIGNAL theory , *COVARIANCE matrices , *BEHAVIOR - Abstract
• NNGPs establish the equivalence between Gaussian processes (GPs) and infinitely wide deep neural networks (NNs). • We consider the impact of noise regularisation (e.g. dropout) on NNGPs using signal propagation theory. • We find that the best NNGPs have kernels matching that of an optimal initialisation for noise regularised ReLU networks. • We also show how noise influences the NNGP's covariance matrix, resulting in simpler posterior functions. • We verify our theoretical findings with experiments on MNIST and CIFAR-10 and synthetic data. Recent work has established the equivalence between deep neural networks and Gaussian processes (GPs), resulting in so-called neural network Gaussian processes (NNGPs). The behaviour of these models depends on the initialisation of the corresponding network. In this work, we consider the impact of noise regularisation (e.g. dropout) on NNGPs, and relate their behaviour to signal propagation theory in noise regularised deep neural networks. For ReLU activations, we find that the best performing NNGPs have kernel parameters that correspond to a recently proposed initialisation scheme for noise regularised ReLU networks. In addition, we show how the noise influences the covariance matrix of the NNGP, producing a stronger prior towards simple functions away from the training points. We verify our theoretical findings with experiments on MNIST and CIFAR-10 as well as on synthetic data. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF