DLCFT: Deep Linear Continual Fine-Tuning for General Incremental Learning
Pre-trained representation is one of the key elements in the success of modern deep learning. However, existing works on continual learning methods have mostly focused on learning models incrementally from scratch. In this paper, we explore an alternative framework to incremental learning where we c...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Pre-trained representation is one of the key elements in the success of
modern deep learning. However, existing works on continual learning methods
have mostly focused on learning models incrementally from scratch. In this
paper, we explore an alternative framework to incremental learning where we
continually fine-tune the model from a pre-trained representation. Our method
takes advantage of linearization technique of a pre-trained neural network for
simple and effective continual learning. We show that this allows us to design
a linear model where quadratic parameter regularization method is placed as the
optimal continual learning policy, and at the same time enjoying the high
performance of neural networks. We also show that the proposed algorithm
enables parameter regularization methods to be applied to class-incremental
problems. Additionally, we provide a theoretical reason why the existing
parameter-space regularization algorithms such as EWC underperform on neural
networks trained with cross-entropy loss. We show that the proposed method can
prevent forgetting while achieving high continual fine-tuning performance on
image classification tasks. To show that our method can be applied to general
continual learning settings, we evaluate our method in data-incremental,
task-incremental, and class-incremental learning problems. |
---|---|
DOI: | 10.48550/arxiv.2208.08112 |