Fine-tuning self-supervised multilingual sequence-to-sequence models for extremely low-resource nmt

dc.contributor.authorThillainathan, S
dc.contributor.authorRanathunga, S
dc.contributor.authorJayasena, S
dc.contributor.editorAdhikariwatte, W
dc.contributor.editorRathnayake, M
dc.contributor.editorHemachandra, K
dc.date.accessioned2022-10-20T04:15:20Z
dc.date.available2022-10-20T04:15:20Z
dc.date.issued2021-07
dc.description.abstractNeural Machine Translation (NMT) tends to perform poorly in low-resource language settings due to the scarcity of parallel data. Instead of relying on inadequate parallel corpora, we can take advantage of monolingual data available in abundance. Training a denoising self-supervised multilingual sequence-to-sequence model by noising the available large scale monolingual corpora is one way to utilize monolingual data. For a pair of languages for which monolingual data is available in such a pre-trained multilingual denoising model, the model can be fine-tuned with a smaller amount of parallel data from this language pair. This paper presents fine-tuning self-supervised multilingual sequence-to-sequence pre-trained models for extremely low-resource domain-specific NMT settings. We choose one such pre-trained model: mBART. We are the first to implement and demonstrate the viability of non-English centric complete fine-tuning on multilingual sequence-to-sequence pre-trained models. We select Sinhala, Tamil and English languages to demonstrate fine-tuning on extremely low-resource settings in the domain of official government documents. Experiments show that our fine-tuned mBART model significantly outperforms state-of-the-art Transformer based NMT models in all pairs in all six bilingual directions, where we report a 4.41 BLEU score increase on Tamil→Sinhala and a 2.85 BLUE increase on Sinhala→ Tamil translation.en_US
dc.identifier.citationS. Thillainathan, S. Ranathunga and S. Jayasena, "Fine-Tuning Self-Supervised Multilingual Sequence-To-Sequence Models for Extremely Low-Resource NMT," 2021 Moratuwa Engineering Research Conference (MERCon), 2021, pp. 432-437, doi: 10.1109/MERCon52712.2021.9525720.en_US
dc.identifier.conferenceMoratuwa Engineering Research Conference 2021en_US
dc.identifier.departmentEngineering Research Unit, University of Moratuwaen_US
dc.identifier.doi10.1109/MERCon52712.2021.9525720en_US
dc.identifier.facultyEngineeringen_US
dc.identifier.pgnos*****en_US
dc.identifier.placeMoratuwa, Sri Lankaen_US
dc.identifier.proceedingProceedings of Moratuwa Engineering Research Conference 2021en_US
dc.identifier.urihttp://dl.lib.uom.lk/handle/123/19156
dc.identifier.year2021en_US
dc.language.isoenen_US
dc.publisherIEEEen_US
dc.relation.urihttps://ieeexplore.ieee.org/document/9525720en_US
dc.subjectneural machine translationen_US
dc.subjectpre-trained modelsen_US
dc.subjectfine-tuningen_US
dc.subjectdenoising autoencoderen_US
dc.subjectlow-resource languagesen_US
dc.titleFine-tuning self-supervised multilingual sequence-to-sequence models for extremely low-resource nmten_US
dc.typeConference-Full-texten_US

Files

Collections