Options
Efficient Chinese-Malay Speech-Text Translation via Layer-Freezing Adaptation of Multimodal Foundation Models
Journal
IEEE Access
ISSN
2169-3536
Date Issued
2025
Author(s)
Xiao Liang
Tien-Ping Tan
Donghong Qin
DOI
10.1109/ACCESS.2025.3568474
Abstract
This paper addresses the challenge of Chinese-Malay speech-to-text translation (S2TT), a crucial yet under-resourced language pair in computational linguistics. We introduce Layer-Freezing Adaptive Fine-Tuning (LFAFT), a parameter-efficient strategy that selectively freezes and unfreezes Transformer layers to optimize model adaptation. LFAFT achieves an 11.8% relative improvement in BLEU-4 scores while reducing trainable parameters by 45% compared to full fine-tuning. Using our newly constructed Chinese-Malay parallel corpus, our approach improves BLEU scores from 1.86 to 9.30 (+7.44 points) compared to existing Chinese-Malay speech translation systems. This work not only establishes the first large-scale Chinese-Malay S2TT dataset but also presents an efficient adaptation method that makes low-resource speech translation more accessible and computationally feasible.
File(s)
Loading...
Name
j.png
Size
17.27 KB
Format
PNG
Checksum
(MD5):85f5e85fa8f8c13d7350540217a227b6
