WaveTransfer: A Flexible End-to-end Multi-instrument Timbre Transfer with Diffusion
2024 IEEE International Workshop on Machine Learning for Signal Processing (MLSP 2024), Sep 2024, London (UK), United Kingdom As diffusion-based deep generative models gain prevalence, researchers are actively investigating their potential applications across various domains, including music synthes...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | 2024 IEEE International Workshop on Machine Learning for Signal
Processing (MLSP 2024), Sep 2024, London (UK), United Kingdom As diffusion-based deep generative models gain prevalence, researchers are
actively investigating their potential applications across various domains,
including music synthesis and style alteration. Within this work, we are
interested in timbre transfer, a process that involves seamlessly altering the
instrumental characteristics of musical pieces while preserving essential
musical elements. This paper introduces WaveTransfer, an end-to-end diffusion
model designed for timbre transfer. We specifically employ the bilateral
denoising diffusion model (BDDM) for noise scheduling search. Our model is
capable of conducting timbre transfer between audio mixtures as well as
individual instruments. Notably, it exhibits versatility in that it
accommodates multiple types of timbre transfer between unique instrument pairs
in a single model, eliminating the need for separate model training for each
pairing. Furthermore, unlike recent works limited to 16 kHz, WaveTransfer can
be trained at various sampling rates, including the industry-standard 44.1 kHz,
a feature of particular interest to the music community. |
---|---|
DOI: | 10.48550/arxiv.2409.15321 |