Bounds on learning in polynomial time

The performance of large neural networks can be judged not only by their storage capacity but also by the time required for learning. A polynomial learning algorithm with learning time α N 2 in a network with N units might be practical whereas a learning time α exp N would allow rather small network...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Philosophical magazine. B, Physics of condensed matter. Structural, electronic, optical, and magnetic properties. Physics of condensed matter. Structural, electronic, optical, and magnetic properties., 1998-05, Vol.77 (5), p.1495-1505
Hauptverfasser: Horner, Heinz, Bethge, Anthea
Format: Artikel
Sprache:eng
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The performance of large neural networks can be judged not only by their storage capacity but also by the time required for learning. A polynomial learning algorithm with learning time α N 2 in a network with N units might be practical whereas a learning time α exp N would allow rather small networks only. The question of absolute storage capacity α c and capacity for polynomial learning rules α p is discussed for several feedforward architectures, the perception, the binary perceptron, the committee machine and a perceptron with fixed weights in the first layer and adaptive weights in the second layer. The analysis is based partially on dynamic mean-field theory which is valid for N → ∞. In particular, for the committee machine a value α p considerably lower than the capacity predicted by replica theory or simulations is found. This discrepancy is resolved by new simulations investigating the learning time dependence and revealing subtleties in the definition of the capacity.
ISSN:1364-2812
0141-8637
1463-6417
DOI:10.1080/13642819808205041