Back to Search Start Over

AudioEditor: A Training-Free Diffusion-Based Audio Editing Framework

Authors :
Jia, Yuhang
Chen, Yang
Zhao, Jinghua
Zhao, Shiwan
Zeng, Wenjia
Chen, Yong
Qin, Yong
Publication Year :
2024

Abstract

Diffusion-based text-to-audio (TTA) generation has made substantial progress, leveraging latent diffusion model (LDM) to produce high-quality, diverse and instruction-relevant audios. However, beyond generation, the task of audio editing remains equally important but has received comparatively little attention. Audio editing tasks face two primary challenges: executing precise edits and preserving the unedited sections. While workflows based on LDMs have effectively addressed these challenges in the field of image processing, similar approaches have been scarcely applied to audio editing. In this paper, we introduce AudioEditor, a training-free audio editing framework built on the pretrained diffusion-based TTA model. AudioEditor incorporates Null-text Inversion and EOT-suppression methods, enabling the model to preserve original audio features while executing accurate edits. Comprehensive objective and subjective experiments validate the effectiveness of AudioEditor in delivering high-quality audio edits. Code and demo can be found at https://github.com/NKU-HLT/AudioEditor.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2409.12466
Document Type :
Working Paper