Back to Search Start Over

TaylorBeamixer: Learning Taylor-Inspired All-Neural Multi-Channel Speech Enhancement from Beam-Space Dictionary Perspective

Authors :
Li, Andong
Yu, Guochen
Liu, Wenzhe
Li, Xiaodong
Zheng, Chengshi
Publication Year :
2022

Abstract

Despite the promising performance of existing frame-wise all-neural beamformers in the speech enhancement field, it remains unclear what the underlying mechanism exists. In this paper, we revisit the beamforming behavior from the beam-space dictionary perspective and formulate it into the learning and mixing of different beam-space components. Based on that, we propose an all-neural beamformer called TaylorBM to simulate Taylor's series expansion operation in which the 0th-order term serves as a spatial filter to conduct the beam mixing, and several high-order terms are tasked with residual noise cancellation for post-processing. The whole system is devised to work in an end-to-end manner. Experiments are conducted on the spatialized LibriSpeech corpus and results show that the proposed approach outperforms existing advanced baselines in terms of evaluation metrics.<br />Comment: In submission to ICASSP 2023, 5 pages

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2211.12024
Document Type :
Working Paper