1. Introduction
Technology reaches new heights through its journey from the origins of ideas to their full-scale practical implementation. One such journey is heading toward the elimination of language barriers in order to establish seamless social communication in every domain.
In this context, advances in related fields such as Natural Language Processing (NLP), Artificial Intelligence-based Language Modeling (LM), and Machine Learning contribute significantly to the development of a faultless automatic machine translation (MT) system. Regardless of the variety of heuristic methods used to preserve both contextual and lexical interpretations, obtaining sufficient adequacy and fluency of the translation model (TM) remains difficult. It requires a large high-quality parallel corpus (translation pairs in source and target languages) (Chapelle and Chung Reference Chapelle and Chung2010). Hence, MT systems prove to be quite efficient with adequate training for high-resource languages, that is, languages with a large quantity of corpora available and having massive digital footprint across the globe (Dorr, Hovy, and Levin Reference Dorr, Hovy and Levin2004). On the contrary, it becomes very complicated for low-resource languages suffering from limited recognition and scanty digital presence. Such imbalance often leads to poor-quality translation. Therefore, MT systems need to understand the syntax (rules to combine words), semantics (meaning of words and combinations), and morphology (rules to cover morphemes—smallest meaningful units—into words) of such low-resource languages adequately (Somers Reference Somers2011).
Statistical MT (SMT) is one of the popular methods proposed to solve these problems. It is a way of translation wherein a statistical-based learning algorithm is applied to a large bilingual corpus that helps the machine learn the translation. This method also enables the machine to translate sentences not encountered by the machine during its training and testing. Unlike rule-based and example-based MT approaches, SMT does not require human intervention, and it can build a translation system directly from training data (Lopez Reference Lopez2008).
On the other hand, Neural MT (NMT) is performed using a neural network (NN) (Stasimioti Reference Stasimioti2020). In contrast to SMT, NMT does not need any TMs, language models, or reordering models. Instead, it makes use of a single sequence model to produce words one at a time. On the basis of the source sentence and the previously formed sequence in the target language, predictions are made. NMT is a deep learning-based technique for MT that operates using word vector representations and a massive NN.
Whether NMT can replace SMT is still debated, with sufficient evidence supporting both sides. Eventually, the experiment of Ziemski, Junczys-Dowmunt, and Pouliquen (Reference Ziemski, Junczys-Dowmunt and Pouliquen2016) on the corpus of the United Nations (consisting of fifteen low-resource languages) emphasizes the fact that SMT and NMT can excel in different situations. From the result of his experiment, it is evident that the performance of SMT is better than that of NMT for the majority of cases, as measured by the BiLingual Evaluation Understudy (BLEU) score. In comparison with SMT, NMT training typically takes a longer time. Research has also shown that when there is a domain incompatibility between testing and training data, SMT performance is superior to that of NMT (Wang, Tu, and Zhang Reference Wang, Tu and Zhang2018; Mahata et al. Reference Mahata, Mandal, Das and Bandyopadhyay2018). Long sentences are another area where SMT excels (Toral and Sánchez-Cartagena Reference Toral and Sánchez-Cartagena2017). Moreover, many researchers (Zhou et al. Reference Zhou, Hu, Zhang and Zong2017; Wang et al. Reference Wang, Lu, Tu, Li, Xiong and Zhang2017; Castilho et al. Reference Castilho, Moorkens, Gaspari and Way2017; Lohar et al. Reference Lohar, Popović, Alfi and Way2019) have pointed out various disadvantages of NMT over SMT on low-resource languages, like Indic languages (ILs), such as the fact that NMT requires more corpus and resources than SMT.
English (EN) and ILs are languages with less parallel data, which motivates us to work with ILs. This research examines the effectiveness of SMT systems on low-resource language pairs, of which many are rarely worked on. The corpus used in the experiment for all fifteen ILs is tested for the first time for building SMT models. These languages are Assamese (AS), Malayalam (ML), Bengali (BN), Marathi (MR), Gujarati (GU), Kannada (KN), Hindi (HI), Oriya (OR), Punjabi (PA), Telugu (TE), Sindhi (SD), Sinhala (SI), Nepali (NE), Tamil (TA), and Urdu (UR).
Our main goal is to develop SMT model for low-resource languages, that is, ILs, which can serve as a baseline system. The following are the summary of our work’s main contributions:
-
This work is the first attempt to build an SMT model with the Samanantar and OPUS Dataset for fifteen IL-EN and EN-IL pairs (both directions), including the Indo-Aryan and Dravidian groups. A realistic assessment of translation quality is evaluated using different evaluation metrics such as BLEU, Metric for Evaluation of Translation with Explicit Ordering (METEOR), Translation Edit Rate (TER), and Rank-based Intuitive Bilingual Evaluation Score (RIBES).
-
Various data filtration methods are investigated in order to clean the data and improve translation quality.
-
A comparison of distance and Morpho-syntactic Descriptor Bidirectional Finite-State Encoder (msd-bidirectional-fe) reordering models is examined to determine the optimal configuration of components in the TMs.
-
SMT models are compared to NMT models to determine how well they perform in translation tasks.
This paper is arranged as follows. Subsection 1.1 gives insight into SMT, while Subsections 1.2 and 1.3 elucidate alignment and reordering models, respectively. In Section 2, prominent works on SMT and NMT using ILs are described. The experimental framework, including an overview of the dataset and methodology with the workflow of the SMT model, is explained in Section 3. Results are presented in Section 4 followed by the conclusion and future work in Section 5.
1.1 SMT
SMT is dependent on statistical methods (Koehn et al. Reference Koehn, Hoang, Birch, Callison-Burch, Federico, Bertoldi, Cowan, Shen, Moran, Zens, Dyer, Bojar, Constantin and Herbst2007; Zens, Och, and Ney Reference Zens, Och and Ney2002; Hearne and Way Reference Hearne and Way2011). It is a data-driven technique that makes use of parallel-aligned corpora. It utilizes Bayesian theory to maximize the likelihood of source-to-target language translation ( $P(Tl|Si)$ , where $Tl$ is the target language and $Si$ is the source input) using equation (1):
SMT is divided into three stages: the LM $P(Tl)$ , the TM $P(Si|Tl)$ , and the decoder model (DM) (Kumawat and Chandra Reference Kumawat and Chandra2014).
The LM relies on the $n$ -gram model to compute the probability of a sentence. It assigns the probability of a single token to the last $n$ tokens that come before it in the sentence and estimates the likelihood of the token. It assumes that the likelihood is independent of any tokens before the last $n$ tokens before it. The chain rule aids in breaking down the sentence into conditional probability products.
where $P(s)$ is the probability of the sentence $s$ , consisting of tokens $w_1$ , $w_2$ , $\ldots$ , $w_n$ , assuming a $k$ -gram model. It utilizes the bilingual parallel corpus of the desired language pair. This is accomplished by calculating the likelihood of tokens extracted from sentences.
The TM is the second phase of SMT, which computes the conditional probability $P(Si \mid Tl)$ , trained from the corpus. The DM is the final and most crucial phase of SMT. It assists in the selection of words with the highest probability to be translated by maximizing the likelihood, that is, $P(Tl) P(Si \mid Tl)$ .
1.2 Alignment model
In MT, an alignment model is used to determine the relationship between phrases or words in the original language and the language being translated. It is important for SMT systems as they help to determine which phrases or words in the source text are associated with particular words or phrases in the destination language.
Giza++ is one such library that is used for alignment. It has different alignment methods such as intersection grow, grow-diag, union, srctotgt, and tgttosrc. Girgzdis et al. (2014) used the Gold Standard as well as the first one million sentences from the DGT-TM 2013 corpus (Steinberger et al. Reference Steinberger, Eisele, Klocek, Pilos and Schlüter2013) and investigated the effect of these alignment models on them. The models were evaluated using the BLEU score. Among all of the methods examined, the “grow-diag-final-and” method received the highest BLEU score, indicating its outstanding efficiency in precisely aligning words and improving translation quality.
The grow-diag-final-and model starts with the intersection of the alignments from source to target and target to source, then two steps are used to add additional alignment points (Costa-jussà et al. Reference Costa-jussà, Cross, Çelebi, Elbayad, Heafield, Heffernan and Kalbassi2003):
-
grow-diag : For every neighboring point to the alignments measured, if either source or target word is not aligned already but is present in the union of the alignment, then the neighboring point is included in the alignment.
-
final : If any phrase pairs are unaligned but present in the union, add the point to the alignment.
It assesses the likelihood of word-to-word alignment for each source and target word in each sentence. To produce a good-quality word alignment, the alignment is produced using a series of successive estimations. The alignment method establishes a connection between the target and source words. This step is followed by reordering.
1.3 Reordering model
Reordering is the process of restructuring the word order of one natural language sentence to make it more similar to the word order of another natural language sentence. It is a critical task in the translation of languages with different syntactic structures. During training, the Moses system determines various reordering possibilities with every phrase. Some of the reordering strategies are msd-bidirectional-fe and distance-based reordering.
1.3.1 msd-bidirectional-fe reordering
msd-bidirectional-fe reordering is an approach utilized during translation to improve the order of words in the target language sentence. It is an improved approach that considers the place of each word within the sentence. It utilizes a set of rules to detect same-side dependencies and bidirectional reordering inside a fixed range, allowing for easier reordering while preserving translation coherence and fluency. This method is especially useful for ILs, which possess complex word orders as a result of their rich morphological and syntactic structure. The advantage of this reordering is that it can provide more accurate and fluent translations by identifying the dependencies within words in both left-to-right and right-to-left directions.
1.3.2 Distance based reordering
Distance reordering is a more straightforward method that reorders words based on their location in the source sentence using a fixed set of rules. It needs to consider the sentence’s syntactic structure or the specific dependencies within words, which otherwise would result in less accurate translations. It assigns a linear cost to reordering distance, implying that the movement of phrases over long distances is more expensive (Kumawat and Chandra Reference Kumawat and Chandra2014). The advantage of distance-based reordering is that it is computationally effective as it distributes reordering costs according to relative distances instead of creating complex structures. However, it may not be as good at capturing long-term interdependence or crucial word order changes as the bidirectional model.
2. Related work
In the year 2009, EN-to-HI SMT system has been created by (Ramanathan et al. Reference Ramanathan, Choudhary, Ghosh and Bhattacharyya2009) using morphological and syntactic preprocessingpre-processing in the SMT model. In their work, the suffixes in HI language are segmented for morphological processing before rearranging the EN source sentences as per HI syntax.
In 2010, research has been conducted by Zbib et al. (Reference Zbib, Matsoukas, Schwartz and Makhoul2010) at MIT, USA, using the grammatical structures in SMT with the Newswire corpus for Arabic to EN language to give better translation results.
Work on KN-to-EN SMT system by Shiva Kumar, Namitha, and Nithya (Reference Shiva Kumar, Namitha and Nithya2015) shows a remarkable feat with 14.5 BLEU score on using Bible corpus on 20,000 sentences. The results are supported by Papineni et al. (Reference Papineni, Roukos, Ward and Zhu2002).
Kaur and Josan (Reference Kaur and Josan2011) presented a TM based on SMT for EN to PA with their own corpus containing 3844 names in both languages with BLEU and word accuracy as 0.4123 (with range 0-1) and 50.22%, respectively.
Nalluri and Kommaluri (Reference Nalluri and Kommaluri2011) has created “enTel,” an SMT-based EN to TE MT system, using the Johns Hopkins University Open Source Architecture (Li et al. Reference Li, Callison-Burch, Dyer, Khudanpur, Schwartz, Thornton, Weese and Zaidan2009). For the purpose of training the translation system, TE parallel dataset from the Enabling Minority Language Engineering (EMILLE) is used for their work.
In the year 2014, an SMT Framework for SI-TA MT System has been created by Pushpananda, Weerasinghe, and Niranjan (Reference Pushpananda, Weerasinghe and Niranjan2014). In their work, the result of SMT-dependent translation between language pairs, including TA-SI and SI-TA has been shown. Outcomes of the experiments using the SMT model give more noticeable results for the SI-TA than the TA-SI language pair. For languages closely related, SMT shows remarkable results.
In 2017, a survey has been conducted by Khan, Anwar, and Durrani (Reference Khan, Anwar and Durrani2017) on the IL-EN language MT models revealing the importance of SMT over eight languages that is, HI, BN, GU, UR, TE, PA, TA, and ML. In their work, EMILLE corpus (Nalluri and Kommaluri Reference Nalluri and Kommaluri2011) is used and Moses SMT model is preferred to make the TMs, with out-of-vocabulary (OOV) words transliterated to EN. In their work, the evaluation using BLEU, NIST and UNK counts as metrics reveals the overall SMT performance as satisfactory (PA-EN and UR-EN models as the best and the HI-EN and GU-EN models as the worst).
An EN-BN SMT system has been presented by Islam, Tiedemann, and Eisele (Reference Islam, Tiedemann and Eisele2010). In their work, a transliteration module is presented to handle OOV words and a preposition handling module has been added to address the systematic grammatical distinctions between EN and BN. BLEU, NIST and TER scores has been used to check the effectiveness of their system.
Nowadays, NMT is widely appreciated for its advancement in the development of MT with remarkable improvement in quality. Hence, many researchers have compared both techniques for low and high-resource languages.
Toral and Sánchez-Cartagena (Reference Toral and Sánchez-Cartagena2017) performed a thorough evaluation using SMT and NMT systems for nine language directions along a variety of dimensions. In their experiment, SMT systems perform better than the NMT for long sentences.
Recently, Castilho et al. (Reference Castilho, Moorkens, Gaspari and Way2017) used automatic metrics and expert translators to conduct a thorough quantitative and qualitative comparison of NMT and SMT. SMT shows better according to their experiments.
The comparison of NMT and SMT for the NE using the Nepali National Corpus with 6535 sentences has been shown by Acharya and Bal (Reference Acharya and Bal2018). The researchers have proved in their experiments that the SMT model performs better than the NMT-based system with a small corpus with a 5.27 BLEU score.
Jiao Liu (Liu Reference Liu2020) provided an extensive comparison and study of the use of cohesive devices by several MT systems in both SMT and NMT circumstances for the Chinese to the EN language. When compared to SMT, the NMT system is more effective in processing (a) additive devices that render EN sentences cohesive, (b) adjectives, (c) and pronouns, indicating better quality of translation.
Abujar et al. (Reference Abujar, Masum, Bhattacharya, Dutta and Hossain2021) developed a BN-EN MT model on AmaderCAT corpus using Sequence-to-Sequence (seq2seq) architecture, a special class of Recurrent Neural Networks to develop the translation system, and has achieved a BLEU score of 22.3.
In the year 2021, translation of EN and HI-to-TA languages using both SMT and NMT has been presented by Ramesh et al. (Reference Ramesh, Parthasarathy, Haque and Way2020). The disadvantages of NMT have been shown in their experiments such as the occurrence of numerous errors by NMT when interpreting domain terms and OOV phrases. NMT frequently constructs inaccurate lexical choices for polysemous words and occasionally counters reordering mistakes while translating words and domain terms. The translations that have been generated by the NMT models mostly include repetitions of previously transcribed words, odd translations, and many unexpected sentences having no correlation with the original sentence.
In the year 2022, WAT2022 (Workshop on Asian Translation 2022) organizes (hosted by the COLING 2022) various works on Indic languages (Nakazawa et al. Reference Nakazawa, Nakayama, Ding, Dabre, Higashiyama, Mino, Goto, Pa, Kunchukuttan, Parida, Bojar, Chu, Eriguchi, Abe, Oda and Kurohashi2022).
Laskar et al. (Reference Laskar, Manna, Pakray and Bandyopadhyay2022) investigated a transliteration-based method in the MNMT model for two ILs, AS and BN, and achieved BLEU scores of 1.10 for EN to AS and 3.50 for AS to EN. One more research paper related to five ILs (Das et al. Reference Das, Biradar, Mishra and Patra2022) also used the MNMT approach with filtration data and achieved an 8.00 BLEU score for NE to the EN language.
3. Experimental framework
This section gives an overview of the dataset and languages used for the experiment and describes the process of building the SMT model.
3.1 Language preference
India is a multilingual nation where people from various states use a variety of regional tongues. Such diversity of language brings difficulty in communicating with one another for information exchange. Further, limitations in public communication also bring inconvenience to share feelings, thoughts, opinions, and facts, as well as to deal with business purposes. Moreover, there are many resources available on the internet in EN but many Indians struggle to take benefit of those due to language barriers (Das et al. Reference Das, Panda, Mishra, Patra and Ekbal2024).
Hence, it is crucial to have an easy translation solution for regional languages to support effective communication and to help utilising global resources. To make it possible, technological innovation is continuing to find out efficient methods for a flawless translation using machines, because it is impractical to have human translators everywhere. For MT, an enormous amount of resources is required for training with a proper knowledge base (rules) for better efficiency so as to fulfill the demand of a flawless translation solution. For translation, understanding the meaning of words is important, but words are not enough to constitute a language as a whole. They must be used in sentence construction that adheres to strict grammar rules and every language is having its own writing style. In the work, fifteen commonly spoken languages (over various regions of India) are chosen. Table 1 describes the languages used for the experiments with their linguistic features (What are the Top 200 Most Spoken Languages ).
3.2 Dataset
OPUS and Samanantar datasets for model building and Flores 200 for fine-tuning and testing are utilized. Samanantar is the largest corpus collection for ILs (Ramesh et al. Reference Ramesh, Doddapaneni, Bheemaraj, Jobanputra, Raghavan, Sharma and Sahoo2022). The collection includes more than 45 million sentence pairs in EN and 11 ILs. The Samanantar Corpus has been used for AS, ML, BN, MR, GU, KN, HI, OR, PA, TE, and TA for the experiments.
OPUS is a large resource with freely available parallel corpus, including data from many domains and covering over 90 languages (Tiedemann Reference Tiedemann2012). The OPUS corpus is used for SI, SD, UR, and NE. Table 2 gives statistics of the dataset used in the experiments.
FLORES-200 (Costa-jussà et al. Reference Costa-jussà, Cross, Çelebi, Elbayad, Heafield, Heffernan and Kalbassi2022) corpora are multilingual parallel corpora with 200 languages, that are used as human-translated benchmarks. They consist of two corpora, labeled “dev” (997 lines) and “devtest” (1,012 lines). The “dev” corpus has been used for fine-tuning, and the “devtest” corpus has been used for testing. These fifteen languages are datasets of WAT2022.
3.3 Methodology
Fig. 1 describes the process of building of the SMT model. The steps for building the SMT model are described below.
-
(1) Pretraining Preprocessing Stage I: Some of the punctuation in the extended Unicode were converted to their standard counterparts. The accented characters were removed. Numbers in the ILs were converted from ENG to Indic scripts. Characters outside the standard alphabets of the language pair, extra spaces, and the unprintable characters were removed from the corpus. The corpus is then tokenized using a modified Moses tokenizer (Koehn et al. Reference Koehn, Hoang, Birch, Callison-Burch, Federico, Bertoldi, Cowan, Shen, Moran, Zens, Dyer, Bojar, Constantin and Herbst2007), which involves dividing a character sequence into smaller tokens, such as words, punctuation, and numerals, based on a given character sequence and a specified document unit. Finally, redundant quotation marks were removed from the corpus. The modified Moses tokenizer is tailored for ILs. It deals with diacritics like halants and nuktas effectively. For example, in AS is modified into .
-
(2) Training Truecasing Model: In Moses, the training procedure utilizes word and segment occurrences to draw connections between the target and source languages. The language and TMs are trained on the training dataset and binarized. GIZA++ grow-diag-final-and alignment is used for word alignments, which start with the intersection of the two alignments and then add the additional alignment points. A truecaser model (a model which changes the words at the beginning of the sentence were changed to the most common casing) is trained on the train dataset. The Moses truecaser is used for the same. The train, dev, and test datasets are truecased using the trained truecase model.
-
(3) Pretraining Preprocessing Stage II: The train, dev, and test datasets are cleaned using clean-corpus-n.perl which removes empty lines and redundant space characters. GIZA++ cannot handle parallel sentences where the ratio of the number of tokens on one side over the other is more than 9:1. Such translations are indicative of too short or too long translations and may indicate incorrect translations. lines that violate the 9:1 sentence ratio limit of GIZA++ are also removed from the corpus.
-
(4) Training Language and TMs: In Moses, the training procedure utilizes word and segment occurrences to draw connections between the target and source languages. The language and TMs are trained on the training dataset and binarized. After training the language model, the next step is to train TMs. The proposed work utilizes the GIZA++ (Costa-jussà et al. Reference Costa-jussà, Cross, Çelebi, Elbayad, Heafield, Heffernan and Kalbassi2003) grow-diag-final-and alignment. Two different reordering methods that is, msd-bidirectional-fe reordering and distance-based reordering are also explored, and their efficiency is checked in terms of ILs in this paper.
-
(5) Fine-tuning: It is the process of determining the best parameters for a TM when it is used for a specific purpose. It uses a TM to translate all fifteen ILs source language phrases in the fine-tuning set. Then, it compares the model’s output to a set of references (human translations) and adjust the variables to improve translation quality. This procedure is repeated several times until the translation quality is optimized. The model is fine-tuned on the preprocessed Flores-200 dev dataset.
-
(6) Translation: The final model is filtered based on the test dataset and then used to translate the preprocessed test dataset from the source to the target language.
-
(7) Postprocessing and Detokenization: Redundant quotation marks were removed, and the translation file is detokenized using the Moses detokenizer.
-
(8) Evaluation: The evaluation metrics used for the experiments are TER (Snover et al. Reference Snover, Dorr, Schwartz, Micciulla and Makhoul2006), METEOR (Banerjee and Lavie Reference Banerjee and Lavie2005), RIBES (Wołk and Koržinek Reference Wołk and Koržinek2016), and BLEU (Papineni et al. Reference Papineni, Roukos, Ward and Zhu2002). Evaluation is done on the Flores-200 devtest dataset. TER can assume any non-negative value, with higher scores being worse. RIBES and METEOR range between 0 and 1 whereas BLEU ranges from 0 to 100 where higher scores reflect better accuracy.
3.4 Neural machine translation (NMT) system
The procedures for NMT are similar to those used for SMT. Initial data processing includes Preprocessing Stage 1, truecasing, preprocessing Stage 2, and then some additional steps such as concatenation of processed data, vocabulary creation and pruning (vocabulary size is 16,000), binarization and model training, detokenization, and finally the calculation of evaluation metrics. The experiment for building the NMT model utilizes the Transformer model (Vaswani et al. Reference Vaswani, Shazeer, Parmar, Uszkoreit, Jones, Gomez, Kaiser and Polosukhin2017) implemented under the Fairseq library (Ott et al. Reference Ott, Edunov, Baevski, Fan, Gross, Ng, Grangier and Auli2019), an open-source sequence modeling toolkit that enables researchers to train custom models for MT tasks. The model comprises six encoder-decoder layers, each with 512 hidden units and optimized with Adam optimizer. The dropout (Srivastava et al. Reference Srivastava, Hinton, Krizhevsky, Sutskever and Salakhutdinov2014) value of 0.3 is applied to each sublayer output before it is added to and normalized with the sublayer input. The model is trained for 300,000 updates to ensure that the training process is completed in a reasonable time frame, balancing model convergence with computational efficiency. The model uses label-smoothed cross entropy as the evaluation criterion. Label smoothing is an effective technique for improving model performance and is a parameter of the label-smoothed cross-entropy loss. It is done so that the model is not overconfident in its prediction and acquires better ways to represent the target language. The maximum number of tokens in a batch is set as 4,096. This series of steps aim to create an NMT model that includes the required preprocessing, training, detokenization, and finally evaluation stages. The best model is used to calculate the evaluation metrics.
4. Results and discussion
Tables 3 and 4 display the evaluation metrics of all the fifteen ILs to EN and vice versa utilizing SMT (with and without fine-tuning) incorporating distance and msd-bidirectional-fe reordering models, respectively.
For EN-IL and IL-EN language using SMT with distance reordering (with and without fine-tuning), the BLEU scores lie between 1.03 to 15.41. The RIBES score lies between 0.08 and 0.63 whereas the METEOR score lies between 0.02 and 0.28. TER ranges between 72.42 and 172.38. From the result, it is observed that the SMT models using the distance reordering techniques are giving better results for languages such as BN, PA, UR, and HI than other languages in terms of all evaluation metrics as shown in Table 3.
It is also noticed that fine-tuning typically decreases the efficacy of the models across all languages, with notable exceptions being SD where fine-tuning helps in increasing efficacy, and AS, where no significant change is observed. Upon further inspection, it is found that the translations present in the two corpora are not accurate. For instance, in the AS corpus, “” is translated as “So if we take some false step before we are aware of it and receive needed counsel from Gods Word, let us imitate Baruchs maturity, spiritual discernment, and humility. Galatians 6: 1.,” but is translated by Google NMT as “( Jeremiah 38:4, 17, 18 ) Like Jeremiah, we declare an unpleasant message to the people.” Similarly, in the SD corpus, “ $\ldots$ ” is literally translated in the dataset as “But fill me with the old familiar Juice,” but is translated by Google NMT as “Tell me, which tribe $\ldots$ .” Therefore, fine-tuning on a more accurate corpus increases the efficacy of the models. For other languages, the difference in linguistic characteristics (differences in the grammar, language structure, or vocabulary) of Samanantar and Flores-200 dev datasets may explain the decrease in efficacy of the models.
Using msd-bidirectional reordering (with and without fine-tuning) the BLEU scores lie between 1.30 and 18.51. TER ranges between 71.21 and 109.37. RIBES scores lie between 0.18 and 0.65, whereas METEOR scores lie between 0.04 and 0.32. For most of the ILs, it is observed that msd-bidirectional reordering performs better than distance reordering. HI, UR, PA, and BN languages have qualitative and less noisy datasets compared to other languages and hence performed better than other languages as shown in Table 4. Additionally, HI, UR, and BN corpora are large in size which contributes to better models. HI performs the best among all languages with and without fine-tuning in all three metrics in both directions in both reordering models, except in IL-EN models with distance reordering and fine-tuning, where PA-EN performs better than the rest. Using both reordering methods, it is observed that ML, SD, and TA perform worse than the other languages, while AS and SI also produce low scores in several cases.
TA and ML are highly agglutinative languages, exhibiting great lexical and morphological diversity. Literature shows that word tokenization does not perform very well in the case of agglutinative languages due to its large and sparse vocabulary. In such cases, techniques such as subword tokenization might be useful.
The inconsistencies among RIBES, METEOR, TER, and BLEU are observed because they concentrate on different aspects of language quality. BLEU is based on the n-gram, which means if the machine-generated output matches the reference text, then it shares more n-gram with the reference text. RIBES scores focus on machine-generated output with good phrase structure and word order. METEOR is intended to detect semantic similarities between the reference and machine-generated output. TER is evaluated by quantifying the edit operations required for a hypothesis to match a reference translation. It is noticed from the result that TER gives inconsistent results as compared to the remaining three metrics. In various cases, the other three metrics indicate that msd bidirectional reordering is better as compared to distance-based reordering, but TER says otherwise. This is because TER only considers a number of edits and does not take into consideration other factors such as location and size of continuous matchings.
It is also noticed that even though the IL-EN and EN-IL systems are trained using the same corpus, the former performs better for all the evaluation metrics. This is due to the significant morphological diversity of ILs and the relative difficulty of translating from EN to ILs. As per our observation using both the reordering technique, performance is comparatively less in the case of small sentences in ILs (such as TA and ML); whereas it is better in the case of languages (like HI and BN) with longer sentences. Table 5 describes the percentile threshold of the number of lines in corpora in 40 percent and 80 percent with respect to the number of tokens in the line.
4.1 Comparison with neural machine translation (NMT) system
Using NMT models BLEU score lies between 0.01 and 17.78, whereas the RIBES score ranges 0.07–0.76. Additionally, the METEOR score exhibits a range of 0.02–0.49, and the RIBES score spans from 61.87 to 592.74 as shown in Table 6. When examining the results for NMT models, it is clear that the AS language consistently gets lower scores across all metrics. In contrast, PA, UR, and BN languages consistently outperform all languages on these metrics.
While comparing evaluation metrics for SMT and NMT models as shown in Table 7, it is observed that there are many cases where either of them is performing better. Metrics are inconsistent with one another such as in many cases BLEU scores evaluate SMT models as better, while the other three metrics show that NMT is better.
5. Conclusion and future work
This paper presented the MT work for fifteen ILs to EN and vice versa using SMT with two different reordering techniques. The grow-diag-final-and alignment model distance and msd-bidirectional-fe reordering models are utilized for the experiment. A tailor-made preprocessing approach has been incorporated, and linguistic features of ILs are elaborated in this work. For checking the quality of translation, different evaluation metrics such as BLEU, RIBES, TER, and METEOR are utilized. From the result, it is observed that the proposed SMT model using the msd-bidirectional-fe reordering technique is better than distance reordering for ILs. However, due to the scarcity and quality of parallel corpus, the metrics obtained are quite low. It has been observed that the translations of the languages are not sufficiently accurate. Measures of validating corpus quality shall be explored in order to observe the corpus quality and remove inaccurate lines. Dravidian languages are in general agglutinative languages (words are made up of morphemes, with each morpheme contributing to the meaning of the word). The paper also includes a comparison between SMT and NMT. The results show that, across various languages, SMT performs better in some cases, while NMT outperforms in others. In the future, techniques to infer translations from the breakdown of words in these languages shall also be explored. Interestingly, in some of the ILs, fine-tuning schemes are hampering the quality. The causes of this phenomenon shall be analyzed and mitigated via techniques such as noise reduction, corpus cleaning, and fine-tuning schemes for those languages to ensure better quality. Other techniques, such as hybridized SMT-NMT models, and the usage of other alignment, agglutination, and reordering models will be studied for further course of research.
Acknowledgment
This research is partially funded by Meity (Ministry of Electronics and Information Technology, Government of India) for project sanction no. 13 (12)/2020-CC BT dated 24.04.2020.