Hostname: page-component-77f85d65b8-5ngxj Total loading time: 0 Render date: 2026-04-18T03:31:09.849Z Has data issue: false hasContentIssue false

Statistical machine translation for Indic languages

Published online by Cambridge University Press:  03 June 2024

Sudhansu Bala Das*
Affiliation:
National Institute of Technology (NIT), Rourkela, Odisha, India
Divyajyoti Panda
Affiliation:
National Institute of Technology (NIT), Rourkela, Odisha, India
Tapas Kumar Mishra
Affiliation:
National Institute of Technology (NIT), Rourkela, Odisha, India
Bidyut Kr. Patra
Affiliation:
Indian Institute of Technology (BHU), Varanasi, Uttar Pradesh, India
*
Corresponding author: Sudhansu Bala Das; Email: baladas.sudhansu@gmail.com
Rights & Permissions [Opens in a new window]

Abstract

Statistical Machine Translation (SMT) systems use various probabilistic and statistical Natural Language Processing (NLP) methods to automatically translate from one language to another language while retaining the originality of the context. This paper aims to discuss the development of bilingual SMT models for translating English into fifteen low-resource Indic languages (ILs) and vice versa. The process to build the SMT model is described and explained using a workflow diagram. Samanantar and OPUS corpus are utilized for training, and Flores200 corpus is used for fine-tuning and testing purposes. The paper also highlights various preprocessing methods used to deal with corpus noise. The Moses open-source SMT toolkit is being investigated for the system’s development. The impact of distance-based reordering and Morpho-syntactic Descriptor Bidirectional Finite-State Encoder (msd-bidirectional-fe) reordering on ILs is compared in the paper. This paper provides a comparison of SMT models with Neural Machine Translation (NMT) for ILs. All the experiments assess the translation quality using standard metrics such as BiLingual Evaluation Understudy, Rank-based Intuitive Bilingual Evaluation Score, Translation Edit Rate, and Metric for Evaluation of Translation with Explicit Ordering. From the result, it is observed that msd-bidirectional-fe reordering performs better than the distance-based reordering model for ILs. It is also noticed that even though the IL-English and English-IL systems are trained using the same corpus, the former performs better for all the evaluation metrics. The comparison between SMT and NMT shows that across various languages, SMT performs better in some cases, while NMT outperforms in others.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press
Figure 0

Table 1. Linguistic features of languages used in MT experiments

Figure 1

Table 2. Parallel corpus statistics

Figure 2

Figure 1. Workflow of Statistical Machine Translation(SMT) model.

Figure 3

Table 3. Evaluation metric result of SMT with and without fine-tuning using distance-based reordering model

Figure 4

Table 4. Evaluation metric result of SMT with and without fine-tuning using msd bidirectional reordering

Figure 5

Table 5. Percentile threshold of corpora

Figure 6

Table 6. Evaluation metrics for Neural Machine Translation (NMT)

Figure 7

Table 7. Translation evaluation metrics by language (right-tick indicates better performance of SMT)