Domain Adaptation for Contrastive Audio-Language Models
Audio-Language Models (ALM) aim to be general-purpose audio models by providing zero-shot capabilities at test time. The zero-shot performance of ALM improves by using suitable text prompts for each domain. The text prompts are usually hand-crafted through an ad-hoc process and lead to a drop in ALM...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Audio-Language Models (ALM) aim to be general-purpose audio models by
providing zero-shot capabilities at test time. The zero-shot performance of ALM
improves by using suitable text prompts for each domain. The text prompts are
usually hand-crafted through an ad-hoc process and lead to a drop in ALM
generalization and out-of-distribution performance. Existing approaches to
improve domain performance, like few-shot learning or fine-tuning, require
access to annotated data and iterations of training. Therefore, we propose a
test-time domain adaptation method for ALMs that does not require access to
annotations. Our method learns a domain vector by enforcing consistency across
augmented views of the testing audio. We extensively evaluate our approach on
12 downstream tasks across domains. With just one example, our domain
adaptation method leads to 3.2% (max 8.4%) average zero-shot performance
improvement. After adaptation, the model still retains the generalization
property of ALMs. |
---|---|
DOI: | 10.48550/arxiv.2402.09585 |