DEEP LEARNING-BASED GURAGIGNA TO AMHARIC MACHINE TRANSLATION

dc.contributor.authorALEMAYEHU BADARGA NIDA
dc.date.accessioned2024-06-18T12:13:49Z
dc.date.available2024-06-18T12:13:49Z
dc.date.issued2024-04
dc.description.abstractMachine translation is an application of NLP, which can be used to translate text from onenatural language to another natural language. In this study, we aimed to develop Deep Learning Based Guragigna to Amharic Translation, recognizing Natural Language Processing as a pivotal domain within AI facilitating human-computer language interaction. Previously, there is no research conducted on machine translation between Guragigna and Amharic. Given the abundance of information in Amharic across various domains in Ethiopia, including legal, media, religious, educational, and governmental documents, it becomes imperative to bridge the language gap for the growing Guragigna-speaking population. Neural Machine Translation (NMT) is a recently proposed approach to machine translation (MT) that has achieved the state-of-the-art translation quality in recent years. Unlike traditional MT approaches, NMT aims to create a single neural network that can be tuned collaboratively to maximize translation performance. So, the aim of this study is to develop Deep learning Amharic-guragigna bi-directional machine translation.To conducted experiments employing six encoder-decoder models: LSTM, Bi-LSTM, LSTM+attention, CNN+attention, GRU and Transformers. Collected a dataset of 9,515 parallel sentences, and evaluated the models based on efficiency metrics, including training time, memory usage, and BLEU score, to propose an optimal translation model and utilize the 80/20 splitting technique for dividing the dataset into training and testing sets. Achieving among those models, the transformer model outperforms other models by 99.4% accuracy, 0.0113 loss and a BLEU score of 9.93 for Amharic-Guragigna translation and 9.99 for Guragigna-Amharic machin translation. Because transformer process the whole sentence simultaneously, which reduces training time and it computes similarity scores between words in a sentence by itself means self attention. Due to the problem of unavailable parallel corpus, we have trained our model with minimum corpus, though NMT requires huge data for training and create an optimal model that learn the different features of the two languages and also challenges with LSTM, Bi-LSTM, LSTM+attention and GRU models, which required significant memory resources. en_US
dc.description.sponsorshipwolkite universtyen_US
dc.identifier.uri
dc.language.isoenen_US
dc.publisherWOLKITE UNIVERSITYen_US
dc.subjectBi-LSTM, CNN+attention, Deep Learning, GRU, LSTM, LSTM+attention, Neural Machine Translation, Transformeren_US
dc.subjectCNN+attentionen_US
dc.subjectDeep Learning,en_US
dc.subjectGRU,en_US
dc.subjectLSTM+attentionen_US
dc.subjectNeural Machine Translation,en_US
dc.subject, Transformeren_US
dc.titleDEEP LEARNING-BASED GURAGIGNA TO AMHARIC MACHINE TRANSLATIONen_US
dc.typeThesisen_US

Files

Original bundle

Now showing 1 - 1 of 1
Thumbnail Image
Name:
Alemayehu Badarga(Deep Learning based Amharic-Guragigna MT).pdf
Size:
3.03 MB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: