Non-deterministic learning dynamics in large neural networks due to structural data bias

H C Rae, J A F Heimel, A C C Coolen

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

We study the dynamics of on-line learning in large (N --> infinity) perceptrons, for the case of training sets with a structural O(N-0) bias of the input vectors, by deriving exact and closed macroscopic dynamical laws using non-equilibrium statistical mechanical tools. In sharp contrast to the more conventional theories developed for homogeneously distributed or only weakly biased data, these laws are found to describe a non-trivial and persistently nondeterministic macroscopic evolution, and a generalization error which retains both stochastic and sample-to-sample fluctuations, even for infinitely large networks. Furthermore, for the standard error-correcting microscopic algorithms (such as the perceptron learning rule) one obtains learning curves with distinct bias-induced phases. Our theoretical predictions find excellent confirmation in numerical simulations.
Original languageEnglish
Pages (from-to)8703 - 8722
Number of pages20
JournalJOURNAL OF PHYSICS A MATHEMATICAL AND GENERAL
Volume33
Issue number48
DOIs
Publication statusPublished - 8 Dec 2000

Fingerprint

Dive into the research topics of 'Non-deterministic learning dynamics in large neural networks due to structural data bias'. Together they form a unique fingerprint.

Cite this