Bidirectional Transformer-Based Neural Machine Translation for Amharic and Tigrinya: Bridging Morphological Complexity and Data Scarcity

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

With the growing reliance on artificial intelligence and digital platforms, which have made more multilingual communication possible. However, languages such as Amharic and Tigrinya, widely spoken in Ethiopia and Eritrea, remain underrepresented in machine translation research due to their complex linguistic structures and limited digital resources. Previous efforts have predominantly employed statistical, rule-based, or hybrid models, which often fail to capture the intricate morphological and syntactic patterns of these languages. In this study, we develop a bidirectional neural machine translation system using a transformer architecture, optimized to accommodate the unique linguistic characteristics of both languages. To address data scarcity, we augmented the training corpus through back translation and employed subword segmentation using SentencePiece to manage morphological richness. We conducted three core experiments: i) word-level tokens on a small original dataset, ii) subword-level modeling, and iii) with augmented data. The final setup yielded the highest performance, achieving BLEU scores of 44.32% (Amharic to Tigrinya) and 44.10% (Tigrinya to Amharic) using greedy decoding, while beam search decoding produced similarly comparable results. These findings highlight the effectiveness of subword tokenization and data augmentation for machine translation in low-resource, morphologically complex languages. Future research may explore morpheme-level representations and advanced decoding techniques to further enhance translation quality.

Article activity feed