Discretization methods for encoding of continuous input variables for Boolean neural networks
RAM-based neural networks are normally based on binary input variables, and a thermometer code or a so-called CMAC-Gray code is most often used when encoding real-valued variables. The number of intervals and interval boundaries are normally set from ad hoc principles. Using this approach many inter...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Tagungsbericht |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | RAM-based neural networks are normally based on binary input variables, and a thermometer code or a so-called CMAC-Gray code is most often used when encoding real-valued variables. The number of intervals and interval boundaries are normally set from ad hoc principles. Using this approach many intervals are normally needed to provide sufficient resolution. This leads to large variable codes, which again complicates the learning problem. Instead of selecting more or less arbitrary interval boundaries if can be expected to be beneficial to use discretization techniques, where the split-values are selected from the use of information measures. We report on the results that can be obtained by applying local and global discretization techniques together with enhanced schemes of the so-called n-tuple classifier which is the most simple type of a RAM neural net. The enhanced n-tuple nets have proven competitive on a large set of benchmark data sets. By making proper use of the discretization boundaries increased performances can be obtained. The local discretization algorithms are closely connected with the learning principle used for decision trees, and we show how such schemes can be used as variable selectors for the RAM based neural nets. |
---|---|
ISSN: | 1098-7576 1558-3902 |
DOI: | 10.1109/IJCNN.1999.831134 |