Graph-based domain adversarial learning framework for video anomaly detection domain generalization

The limited domain generalization capability of contemporary video anomaly detection methods restricts their efficacy to specific datasets. To enhance the generalizability and portability of video anomaly detection models, we propose a domain adaptation network framework with robust generalization p...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:The Journal of supercomputing 2024-09, Vol.80 (13), p.18977-19002
Hauptverfasser: Mei, Xue, Wei, Yachuan, Chen, Haoyang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The limited domain generalization capability of contemporary video anomaly detection methods restricts their efficacy to specific datasets. To enhance the generalizability and portability of video anomaly detection models, we propose a domain adaptation network framework with robust generalization performance. The objective of the framework is to enable the video anomaly detection model to generalize from the source domain to the untrained target domain while mitigating the impact of missing labeled data on deep architectures. The framework incorporates a graph-based domain-invariant representation learning module and domain discriminator that enable the model to learn deep features with domain-invariant properties that remain unchanged across different domains by calculating the strength of the relationships among domain nodes. Notably, inspired by domain adversarial learning, the framework utilizes a gradient reversal layer acting on backpropagation that guides the parameters of optimal feature mapping in constructing the loss with opposing directions. To address the domain generalization problem in video anomaly detection, this framework applies graph convolution techniques. The framework leverages a novel adjacency matrix that encourages high coherence within the same domain while optimizing the mapping of low-level deep features from source to target domains to enhance the discriminative performance of the video anomaly detection model in the target domain. Simulation experiments were conducted on Avenue, UCSD-Ped1, UCSD-Ped2, ShanghaiTech, UCF-Crime, and TAD datasets, and labeled data from the source domain were utilized during the training process. Various testing results demonstrate that our framework enables models trained in one or more different scenes (domains) to perform well in unknown scenes (domains) with good cross-domain testing AUC performance. For example, in multidomain training generalization to the Avenue dataset for testing, our domain adversarial learning framework improves detection accuracy by 12.47%. Under severe single-domain generalization scenarios, the AUC performance on the target domain (e.g., UCF-Crime dataset) increase by 4.36%, 8.64%, and 3.68%, respectively.
ISSN:0920-8542
1573-0484
DOI:10.1007/s11227-024-06154-1