A comparative study of natural language inference in Swahili using monolingual and multilingual models
Abstract
Recent advancements in large language models (LLMs) have led to opportunities for improving applications across various domains. However, existing LLMs fine-tuned for Swahili or other African languages often rely on pre-trained multilingual models, resulting in a relatively small portion of training data dedicated to Swahili. In this study, we compare the performance of monolingual and multilingual models in Swahili natural language inference tasks using the cross-lingual natural language inference (XNLI) dataset. Our research demonstrates the superior effectiveness of dedicated Swahili monolingual models, achieving an accuracy rate of 69%. These monolingual models exhibit significantly enhanced precision, recall, and F1 scores, particularly in predicting contradiction and neutrality. Overall, the findings in this article emphasize the critical importance of using monolingual models in low-resource language processing contexts, providing valuable insights for developing more efficient and tailored natural language processing systems that benefit languages facing similar resource constraints.
Keywords
Monolingual model; Multilingual model; Natural language inference; Swahili; XNLI dataset
Full Text:
PDFDOI: http://doi.org/10.11591/ijai.v14.i2.pp1597-1604
Refbacks
- There are currently no refbacks.
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
IAES International Journal of Artificial Intelligence (IJ-AI)
ISSN/e-ISSN 2089-4872/2252-8938
This journal is published by the Institute of Advanced Engineering and Science (IAES) in collaboration with Intelektual Pustaka Media Utama (IPMU).