Learning a Consensus Sub-Network with Polarization Regularization and One Pass Training
The subject of green AI has been gaining attention within the deep learning community given the recent trend of ever larger and more complex neural network models. Existing solutions for reducing the computational load of training at inference time usually involve pruning the network parameters. Pru...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The subject of green AI has been gaining attention within the deep learning
community given the recent trend of ever larger and more complex neural network
models. Existing solutions for reducing the computational load of training at
inference time usually involve pruning the network parameters. Pruning schemes
often create extra overhead either by iterative training and fine-tuning for
static pruning or repeated computation of a dynamic pruning graph. We propose a
new parameter pruning strategy for learning a lighter-weight sub-network that
minimizes the energy cost while maintaining comparable performance to the fully
parameterised network on given downstream tasks. Our proposed pruning scheme is
green-oriented, as it only requires a one-off training to discover the optimal
static sub-networks by dynamic pruning methods. The pruning scheme consists of
a binary gating module and a polarizing loss function to uncover sub-networks
with user-defined sparsity. Our method enables pruning and training
simultaneously, which saves energy in both the training and inference phases
and avoids extra computational overhead from gating modules at inference time.
Our results on CIFAR-10, CIFAR-100, and Tiny Imagenet suggest that our scheme
can remove 50% of connections in deep networks with |
---|---|
DOI: | 10.48550/arxiv.2302.10798 |