Support subsets estimation for support vector machines retraining

•A new retraining methodology for SVMs is proposed•The new concept of Support Subsets is introduced•Methodology proposal reduces complexity computation for SVM training•Imbalance datasets produce balanced Support Subsets•The proposal compares successfully to well-known retraining techniques The avai...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Pattern recognition 2023-02, Vol.134, p.109117, Article 109117
Hauptverfasser: Aceña, Víctor, Martín de Diego, Isaac, R. Fernández, Rubén, M. Moguerza, Javier
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:•A new retraining methodology for SVMs is proposed•The new concept of Support Subsets is introduced•Methodology proposal reduces complexity computation for SVM training•Imbalance datasets produce balanced Support Subsets•The proposal compares successfully to well-known retraining techniques The availability of new data in previously trained Machine Learning (ML) models usually requires retraining and adjustment of the model. Support Vector Machines (SVMs) are widely used in ML because of their strong mathematical foundations and flexibility. However, SVM training is computationally expensive, both in time and memory. Hence, the training phase might be a limitation in problems where the model is updated regularly. As a solution, new methods for training and updating SVMs have been proposed in the past. In this paper, we introduce the concept of Support Subset and a new retraining methodology for SVMs. A Support Subset is a subset of the training set, such that retraining a ML model with this subset and the new data is equivalent to training with all the data. The performance of the proposal is evaluated in a variety of experiments on simulated and real datasets in terms of time, quality of the solution, resultant support vectors, and amount of employed data. The promising results provide a new research line for improving the effectiveness and adaptability of the proposed technique, including its generalization to other ML models.
ISSN:0031-3203
1873-5142
DOI:10.1016/j.patcog.2022.109117