Back to Search Start Over

Towards Inducing Long-Context Abilities in Multilingual Neural Machine Translation Models

Authors :
Gumma, Varun
Chitale, Pranjal A.
Bali, Kalika
Publication Year :
2024

Abstract

Neural Machine Translation (NMT) models have traditionally used Sinusoidal Positional Embeddings (PEs), which often struggle to capture long-range dependencies and are inefficient for handling extended context or document-level translation tasks. This work addresses the challenge of transitioning pre-trained NMT models from absolute Sinusoidal PEs to Relative PEs, such as RoPE and ALiBi, without compromising performance. We demonstrate that parameter-efficient fine-tuning, using only a small amount of high-quality data, can successfully facilitate this transition. Experimental results indicate that switching from Sinusoidal to Relative PEs results in competitive translation quality on sentence-level evaluation benchmarks. Additionally, models trained with RoPE consistently outperform those using ALiBi and Sinusoidal PEs on document-level benchmarks across both string-based metrics and qualitative evaluations. Moreover, we find that a small amount of long-context data in a few languages is sufficient for cross-lingual length generalization, thereby inducing long-context capabilities.<br />Comment: Accepted at NAACL 2025

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2408.11382
Document Type :
Working Paper