Adversarial examples in Arabic language

Safae Laatyaoui, Mohammed Saber

Abstract


Adversarial attacks have a great popularity in the artificial intelligence (AI) domain. In the natural language processing (NLP) field, various techniques have been used to evaluate the vulnerability of deep learning (DL) models. It is observed that while most studies focused on generating adversarial examples in English language, Arabic adversarial attacks have received little attention. This paper presents a two-step method to create adversarial examples in Arabic language: first, the most important words are identified. Then, the proposed transformation algorithm is applied. Only small and imperceptible manipulations based on common mistakes in Arabic writing mislead the popular pre-trained language model (PLM) bidirectional encoder representations from transformers (BERT) retrained on the book reviews in Arabic dataset (BRAD) on the sentiment analysis (SA) task and decrease its performance: the classification accuracy was reduced by an average of 3.44%. This drop in accuracy shows that the model was successfully attacked.

Keywords


Adversarial attack; BERT; Low-resource language; Natural language processing; Pre-trained language model; Sentiment analysis

Full Text:

PDF


DOI: http://doi.org/10.11591/ijai.v15.i1.pp941-952

Refbacks

  • There are currently no refbacks.


Copyright (c) 2026 Safae Laatyaoui, Mohammed Saber

Creative Commons License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.

IAES International Journal of Artificial Intelligence (IJ-AI)
ISSN/e-ISSN 2089-4872/2252-8938 
This journal is published by the Institute of Advanced Engineering and Science (IAES).

View IJAI Stats