-
Exploring protein profiles and hub genes in ameloblastom
- Back
Metadata
Document Title
Exploring protein profiles and hub genes in ameloblastom
Author
Sanguansin S., Kengkarn S., Klongnoi B., Chujan S., Roytrakul S., Kitkumthorn N.
Affiliations
School of ICT, Sirindhorn International Institute of Technology, Thammasat University, 131 Moo 5 Tiwanon Road, Bangkadi, Mueang Pathum Thani District, Pathum Thani, 12000, Thailand; Language and Semantic Technology Research Team (LST), Artificial Intelligence Research Group (AINRG), National Electronics and Computer Technology Center (NECTEC), 112 Phahonyothin Road, Khlong Nueng, Khlong Luang District, Pathumthani, 12120, Thailand; Laboratory for Future Interdisciplinary Research of Science and Technology, Tokyo Institute of Technology, 4259 Nagatsuta, Midori-ku, Yokohama, Japan
Type
Article
Source Title
ACM Transactions on Asian and Low-Resource Language Information Processing
ISSN
23754699
Year
2024
Volume
23
Issue
4
Open Access
All Open Access, Hybrid Gold
Publisher
Association for Computing Machinery
DOI
10.1145/3645111
Abstract
Several methodologies have recently been proposed to enhance the performance of low-resource Neural Machine Translation (NMT). However, these techniques have yet to be explored thoroughly in the low-resource Thai and Myanmar languages. Therefore, we first applied augmentation techniques such as SwitchOut and Ciphertext Based Data Augmentation (CipherDAug) to improve NMT performance in these languages. Second, we enhanced the NMT performance by fine-Tuning the pre-Trained Multilingual Denoising BART model (mBART), where BART denotes Bidirectional and Auto-Regressive Transformer. We implemented three NMT systems: namely, Transformer+SwitchOut, Multi-Source Transformer+CipherDAug, and fine-Tuned mBART in the bidirectional translations of Thai-English-Myanmar language pairs from the ASEAN-MT corpus. Experimental results showed that Multi-Source Transformer+CipherDAug significantly improved Bilingual Evaluation Understudy (BLEU), Character n-gram F-score (ChrF), and Translation Error Rate (TER) scores over the first baseline Transformer and second baseline Edit-Based Transformer. The model achieved notable BLEU scores: 37.9 (English-To-Thai), 42.7 (Thai-To-English), 28.9 (English-To-Myanmar), 31.2 (Myanmar-To-English), 25.3 (Thai-To-Myanmar), and 25.5 (Myanmar-To-Thai). The fine-Tuned mBART model also considerably outperformed the two baselines, except for the Myanmar-To-English pair. SwitchOut improved over the second baseline in all pairs and performed similarly to the first baseline in most cases. Last, we performed detailed analyses verifying that the CipherDAug and mBART models potentially facilitate improving low-resource NMT performance in Thai and Myanmar languages. ? 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM.
License
CC BY
Rights
Authors
Publication Source
WoS