Back to Search Start Over

Ideal-LLM: Integrating Dual Encoders and Language-Adapted LLM for Multilingual Speech-to-Text

Authors :
Xue, Hongfei
Ren, Wei
Geng, Xuelong
Wei, Kun
Li, Longhao
Shao, Qijie
Yang, Linju
Diao, Kai
Xie, Lei
Publication Year :
2024

Abstract

Integrating audio encoders with LLMs through connectors has enabled these models to process and comprehend audio modalities, significantly enhancing speech-to-text tasks, including automatic speech recognition (ASR) and automatic speech translation (AST). However, these methods often overlook the critical aspect of language adaptation in multilingual settings, relying instead on multilingual data without adequately addressing language differences. To address this gap, we propose the Ideal-LLM model, which employs dual multilingual encoders to enrich language feature information and utilizes a language-adapted connector to target the adaptation of each language specifically. By leveraging the complementary strengths of Whisper and MMS encoders, our approach ensures richer multilingual representations. Additionally, the language-adapted connector enhances modal transformation via a language weight selector tailored for each language. Experimental results demonstrate that Ideal-LLM significantly improves ASR performance, achieving a 32.6% relative reduction in average word error rates compared to the standard speech encoder integrated with LLMs and yields an average BLEU score of 36.78 for AST task.<br />Comment: 5 pages, 3 figures, submitted to ICASSP 2025

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2409.11214
Document Type :
Working Paper