Bounds on learning in polynomial time
The performance of large neural networks can be judged not only by their storage capacity but also by the time required for learning. A polynomial learning algorithm with learning time α N 2 in a network with N units might be practical whereas a learning time α exp N would allow rather small network...
Gespeichert in:
Veröffentlicht in: | Philosophical magazine. B, Physics of condensed matter. Structural, electronic, optical, and magnetic properties. Physics of condensed matter. Structural, electronic, optical, and magnetic properties., 1998-05, Vol.77 (5), p.1495-1505 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The performance of large neural networks can be judged not only by their storage capacity but also by the time required for learning. A polynomial learning algorithm with learning time α N
2
in a network with N units might be practical whereas a learning time α exp N would allow rather small networks only. The question of absolute storage capacity α
c
and capacity for polynomial learning rules α
p
is discussed for several feedforward architectures, the perception, the binary perceptron, the committee machine and a perceptron with fixed weights in the first layer and adaptive weights in the second layer. The analysis is based partially on dynamic mean-field theory which is valid for N → ∞. In particular, for the committee machine a value α
p
considerably lower than the capacity predicted by replica theory or simulations is found. This discrepancy is resolved by new simulations investigating the learning time dependence and revealing subtleties in the definition of the capacity. |
---|---|
ISSN: | 1364-2812 0141-8637 1463-6417 |
DOI: | 10.1080/13642819808205041 |