Back to Search
Start Over
Modular Speech-to-Text Translation for Zero-Shot Cross-Modal Transfer
- Source :
- Proceedings of Interspeech 2023
- Publication Year :
- 2023
-
Abstract
- Recent research has shown that independently trained encoders and decoders, combined through a shared fixed-size representation, can achieve competitive performance in speech-to-text translation. In this work, we show that this type of approach can be further improved with multilingual training. We observe significant improvements in zero-shot cross-modal speech translation, even outperforming a supervised approach based on XLSR for several languages.
- Subjects :
- Computer Science - Computation and Language
Subjects
Details
- Database :
- arXiv
- Journal :
- Proceedings of Interspeech 2023
- Publication Type :
- Report
- Accession number :
- edsarx.2310.03724
- Document Type :
- Working Paper
- Full Text :
- https://doi.org/10.21437/Interspeech.2023-2484