1. Energy Conservation in Infinitely Wide Neural-Networks
- Author
-
Shu Eguchi and Takafumi Amaba
- Subjects
Quantitative Biology::Neurons and Cognition ,Artificial neural network ,Stochastic process ,sort ,Statistical physics ,Type (model theory) ,Layer (object-oriented design) ,Conserved quantity ,Parametrization ,Energy (signal processing) ,Mathematics - Abstract
A three-layered neural-network (NN), which consists of an input layer, a wide hidden layer and an output layer, has three types of parameters. Two of them are pre-neuronal, namely, thresholds and weights to be applied to input data. The rest is post-neuronal weights to be applied after activation. The current paper consists of the following two parts. First, we consider three types of stochastic processes. They are constructed by summing up each of parameters over all neurons at each epoch, respectively. The neuron number will be regarded as another time different to epochs. In the wide neural-network with a neural-tangent-kernel- (NTK-) parametrization, it is well known that these parameters are hardly varied from their initial values during learning. We show that, however, the stochastic process associated with the post-neuronal parameters is actually varied during the learning while the stochastic processes associated with the pre-neuronal parameters are not. By our result, we can distinguish the type of parameters by focusing on those stochastic processes. Second, we show that the variance (sort of “energy”) of the parameters in the infinitely wide neural-network is conserved during the learning, and thus it gives a conserved quantity in learning.
- Published
- 2021