Back to Search Start Over

On learning and energy-entropy dependence in recurrent and nonrecurrent signed networks

Authors :
Stephen Grossberg
Source :
Journal of Statistical Physics. 1:319-350
Publication Year :
1969
Publisher :
Springer Science and Business Media LLC, 1969.

Abstract

Learning of patterns by neural networks obeying general rules of sensory transduction and of converting membrane potentials to spiking frequencies is considered. Any finite number of cellsA can sample a pattern playing on any finite number of cells ∇ without causing irrevocable sampling bias ifA = ℬ orA ∩ ℬ = . Total energy transfer from inputs ofA to outputs of ℬ depends on the entropy of the input distribution. Pattern completion on recall trials can occur without destroying perfect memory even ifA = ℬ by choosing the signal thresholds sufficiently large. The mathematical results are global limit and oscillation theorems for a class of nonlinear functional-differential systems.

Details

ISSN :
15729613 and 00224715
Volume :
1
Database :
OpenAIRE
Journal :
Journal of Statistical Physics
Accession number :
edsair.doi...........7712118149837b18a2f109a4b73bfa58
Full Text :
https://doi.org/10.1007/bf01007484