Unsupervised bin-wise pre-training: A fusion of information theory and hypergraph

Pre-training is considered to be a ‘triggering point’ for Deep Neural Networks which gains attraction of the researchers. Though recent research works focus on designing efficient pre-training models, they often fail to capture the relevant information representations across the layers with minimum...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Knowledge-based systems 2020-05, Vol.195, p.105650, Article 105650
Hauptverfasser: Glory, H Anila, Vigneswaran, C, Sriram, VS Shankar
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Pre-training is considered to be a ‘triggering point’ for Deep Neural Networks which gains attraction of the researchers. Though recent research works focus on designing efficient pre-training models, they often fail to capture the relevant information representations across the layers with minimum turns and to maintain the stability of the learning model. This research article presents a novel unsupervised bin-wise pre-training model which fuses Information Theory and Hypergraph that acts as an effective optimizer: speed up the learning process & minimize generalization loss, and also as an impelling regularizer: maintain the stability of the Deep Neural Network. The proposed model is evaluated using three different benchmark datasets and the experimental results confirm the supremacy of the proposed model over the state-of-the-art approaches in speeding up the learning process and minimizing generalization loss without deteriorating the stability of Deep Neural Network. •Novel pre-training model is proposed to improve generalization & rate of convergence.•New parameter updation is introduced that performs both optimization & regularization•K-helly property of hypergraph is employed to restraint updation during pre-training.•Three benchmark datasets are used to evaluate the supremacy of the proposed model.
ISSN:0950-7051
1872-7409
DOI:10.1016/j.knosys.2020.105650