Contrastive Deep Supervision
The success of deep learning is usually accompanied by the growth in neural network depth. However, the traditional training method only supervises the neural network at its last layer and propagates the supervision layer-by-layer, which leads to hardship in optimizing the intermediate layers. Recen...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The success of deep learning is usually accompanied by the growth in neural
network depth. However, the traditional training method only supervises the
neural network at its last layer and propagates the supervision layer-by-layer,
which leads to hardship in optimizing the intermediate layers. Recently, deep
supervision has been proposed to add auxiliary classifiers to the intermediate
layers of deep neural networks. By optimizing these auxiliary classifiers with
the supervised task loss, the supervision can be applied to the shallow layers
directly. However, deep supervision conflicts with the well-known observation
that the shallow layers learn low-level features instead of task-biased
high-level semantic features. To address this issue, this paper proposes a
novel training framework named Contrastive Deep Supervision, which supervises
the intermediate layers with augmentation-based contrastive learning.
Experimental results on nine popular datasets with eleven models demonstrate
its effects on general image classification, fine-grained image classification
and object detection in supervised learning, semi-supervised learning and
knowledge distillation. Codes have been released in Github. |
---|---|
DOI: | 10.48550/arxiv.2207.05306 |