Please use this identifier to cite or link to this item:
https://elib.vku.udn.vn/handle/123456789/6235| Title: | Fine-Tuning Multilingual Khmer Neural Machine Translation |
| Authors: | Rina, Buoy Sovisal, Chenda Nguonly, Taing Marry, Kong Masakazu, Iwamura Koichi, Kise |
| Keywords: | Khmer neural machine translation No language left behind (NLLB) Low-resource languages |
| Issue Date: | Jan-2026 |
| Publisher: | Springer Nature |
| Abstract: | Google Translate remains a strong baseline machine translation (MT) tool for Khmer. However, as a proprietary tool, it does not allow flexible deployment, customization, or improvement. In contrast, “No Language Left Behind” (NLLB) is an open-source MT solution, but its translation performance for Khmer is significantly weaker than that of Google Translate. Given the low-resource nature of the Khmer language, this paper pragmatically presents a robust machine translation model for translating Khmer to and from English, Thai, Vietnamese, and Laotian. This model is developed by fine-tuning a base NLLB model on a high-quality multilingual parallel corpus. The fine-tuned model achieves performance competitive to Google Translate while significantly outperforming the base NLLB model and the previous studies. |
| Description: | Lecture Notes in Networks and Systems (LNNS,volume 1581); The 14th Conference on Information Technology and Its Applications (CITA 2025) ; pp: 87-99 |
| URI: | https://doi.org/10.1007/978-3-032-00972-2_7 https://elib.vku.udn.vn/handle/123456789/6235 |
| ISBN: | 978-3-032-00971-5 (p) 978-3-032-00972-2 (e) |
| Appears in Collections: | CITA 2025 (International) |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.