Please use this identifier to cite or link to this item: https://elib.vku.udn.vn/handle/123456789/6235
Title: Fine-Tuning Multilingual Khmer Neural Machine Translation
Authors: Rina, Buoy
Sovisal, Chenda
Nguonly, Taing
Marry, Kong
Masakazu, Iwamura
Koichi, Kise
Keywords: Khmer neural machine translation
No language left behind (NLLB)
Low-resource languages
Issue Date: Jan-2026
Publisher: Springer Nature
Abstract: Google Translate remains a strong baseline machine translation (MT) tool for Khmer. However, as a proprietary tool, it does not allow flexible deployment, customization, or improvement. In contrast, “No Language Left Behind” (NLLB) is an open-source MT solution, but its translation performance for Khmer is significantly weaker than that of Google Translate. Given the low-resource nature of the Khmer language, this paper pragmatically presents a robust machine translation model for translating Khmer to and from English, Thai, Vietnamese, and Laotian. This model is developed by fine-tuning a base NLLB model on a high-quality multilingual parallel corpus. The fine-tuned model achieves performance competitive to Google Translate while significantly outperforming the base NLLB model and the previous studies.
Description: Lecture Notes in Networks and Systems (LNNS,volume 1581); The 14th Conference on Information Technology and Its Applications (CITA 2025) ; pp: 87-99
URI: https://doi.org/10.1007/978-3-032-00972-2_7
https://elib.vku.udn.vn/handle/123456789/6235
ISBN: 978-3-032-00971-5 (p)
978-3-032-00972-2 (e)
Appears in Collections:CITA 2025 (International)

Files in This Item:

 Sign in to read



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.