pretraining in combination with this ground-breaking
learning technique for low-resource languages.
Thillainathan et al. (Thillainathan, Ranathunga,
et al. , 2021) examines enhancing Accurate
Translation of Low Resource Languages employing
mBART or other related NMT pre-trained models.
The study introduces translation from and to Sinhala
and Tamil and shows by fine-tuning mBART with
little parallel data (e.g., 66,000 sentences), we can
achieve substantial BLEU gains over a comparable
transformer-based NMT model. According to the
findings, the quality of translation is significant on the
amount of monolingual corpus for the target language
and the linguistic density of the language in question.
This research proves that the power of multilingual
models can be effective in the extreme low-resource
setting further implying that the research direction
can proceed toward joint multilingual finetuning or
using even more advanced models such as mT5.
Tran et al. (Tran, 2024) explore ways by which we
can obtain good translations between low resource
language pairs including Lao -Vietnamese. Based on
a dataset of the VLSP 2023 MT challenge, the study
investigates hyperparameters tuning, back
translation, and fine-tuning of multilingual pre-
trained models that include mT5 and mBART. From
the experiments, it can be seen that hyperparameters
tuning yields 22 more BLEU points than experiment
without tuning, back translation increases scores to
27.79 and fine tuning mT5 got the highest score of
28.05. The results show that integrating optimization
with the application of pre-trained models
significantly improve the translations and future work
on low-resource languages.
Hallac et al. (Hallac, Ay, et al. , 2018) further
investigates pretraining and finetuning of deep
learning models for the classification of tweet data
using a large corpus of news articles labeled for the
same topic and a small set of tweets. The authors
employ models such as CNN, Bi-LSTM-CONV, and
MLP first on news data and then fine-tuning them on
tweets to categorise content into culture, economy,
politics, sports, and technology. Altogether, the
experimental evaluation indicate that the fine-tuned
model that performs the best is the Bi-LSTM-CONV
model with high extra accuracy beyond the models
trained solely with tweets. The study implies that the
classification of texts could be improved during pre-
training on similar large datasets and activation of
step-by-step fine-tuning in data-deficient
environments.
Saji et al. (Saji, Chandran, Alharbi, 2022)
discusses an architecture of English-to-Malayalam
machine translation exploiting transformers while
emphasizing translation quality enhancement to low-
resource languages such as Malayalam. It compares
multiple architectures of NMT: Seq2Seq models with
Bahdanau, multi-head and scaled dot product
attention mechanisms, and MarianMT. Adjustment of
the MarianMT model considerably improves
performance, and the solutions obtained have the
highest BLEU and E-values with subjective
estimations. The work also shows that attention
mechanisms help in the enhancement of translation
quality and indicates how these models can be used
in low-resource languages.
Premjith et al. (Premjith, Kumar, et al. , 2019)
The study introduces a Neural Machine Translation
(NMT) system that uses parallel corpora to translate
English into four Indian languages: Tamil, Punjabi,
Hindi, and Malayalam. It draws attention to issues
like the dearth of high-quality datasets and the
morphological diversity of Indian languages, and it
suggests solutions including transliteration modules
to handle terms that are not in the vocabulary and
attention mechanisms for processing lengthy phrases.
Nair et al. (Nair, Krishnan, et al. , 2016) In order
to handle grammatical subtleties like declensions and
sentence reordering, the study suggests a hybrid
strategy for an English-to-Hindi machine translation
system that combines rule-based and statistical
techniques. Its potential for more extensive
multilingual applications is shown by its better
accuracy as compared to current systems.
Unnikrishnan et al. In order to overcome
linguistic disparities, the study presents a Statistical
Machine Translation (SMT) system for English to
South Dravidian languages (Malayalam and
Kannada). It incorporates morphological information,
syntax reordering, and optimized bilingual corpus
construction. It offers a framework that may be
modified to accommodate other Dravidian languages
and exhibits increased translation accuracy and a
smaller corpus size.
KM et al (KM, Namitha, et al. , 2015) In this
paper, two different corpora—a general text corpus
and a Bible text corpus—are used to compare
English-to-Kannada statistical machine translation
(SMT). The difficulties presented by Kannada’s
morphological diversity are emphasized, and
methods for boosting translation quality are covered,
with a focus on how corpus size and token frequency
might raise the baseline SMT systems’ BLEU score.
The next section explains the methodology of our
proposed fine-tuned models.