Few-Shot Object Counting with Dynamic Similarity-Aware in Latent Space

Few-shot object counting (FSOC) estimates object quantities in query images using a few of support information. Unlike traditional counting methods, FSOC prioritizes more discriminative and generalized similarity measures between query and support data. This facilitates counting objects from new cat...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on geoscience and remote sensing 2024-01, Vol.62, p.1-1
Hauptverfasser: He, Jinghui, Liu, Bo, Cao, Fan, Xu, Jian, Xiao, Yanshan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Few-shot object counting (FSOC) estimates object quantities in query images using a few of support information. Unlike traditional counting methods, FSOC prioritizes more discriminative and generalized similarity measures between query and support data. This facilitates counting objects from new categories without extensive datasets creation or costly retraining. However, existing approaches often rely on fixed similarity rules, leading to spatial information loss. Limited training data can yield sparse similarity features distribution, hampering the model's learning and its ability to handle objects with large intra-class difference. In this study, we introduce a novel FSOC network named DSALVANet, which comprising the Dynamic Similarity-Aware module (DSAM) and the Latent Variable Augmentation module (LVAM). DSAM establishes adaptive metric rules for support features to find similar regions in the metric space for accurate object counting. LVAM utilizes prior similarity knowledge from DSAM to modeling the latent distribution of densitiy map, improving the decoder's robustness by sampling diverse latent variables during training. Extensive experiments on FSOC benchmark and remote sensing datasets demonstrate our method's effectiveness and state-of-the-art performance. The code and model are available at DSALVANet.
ISSN:0196-2892
1558-0644
DOI:10.1109/TGRS.2024.3350383