A novel LightGBM model for Arabic spam detection integrated with XAI for enhanced explainability

Arabic spam detection presents technical challenges due to linguistic variability, feature sparsity, and the limited availability of interpretable classification systems. Several machine learning models lack intrinsic interpretability, which reduces transparency in decision-making. This study propos...

Full description

Saved in:
Bibliographic Details
Main Authors: Bouke, Mohamed Aly, Alramli, Omar Imhemed, Abdalhafid, Alsadg Ahmed Albadwi, Abdullah, Azizol
Format: Article
Language:en
Published: Elsevier 2026
Subjects:
Online Access:http://psasir.upm.edu.my/id/eprint/123346/1/123346.pdf
http://psasir.upm.edu.my/id/eprint/123346/
https://www.sciencedirect.com/science/article/pii/S0045790626001047
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Arabic spam detection presents technical challenges due to linguistic variability, feature sparsity, and the limited availability of interpretable classification systems. Several machine learning models lack intrinsic interpretability, which reduces transparency in decision-making. This study proposes a spam detection pipeline that combines LightGBM with an integrated explainability layer. Interpretability is incorporated directly into the evaluation process by embedding SHAP and LIME to quantify explanation behavior. The system targets characteristics of Arabic text data by operating in a compressed feature space. From an initial 100-dimensional TF-IDF representation, the top 20 statistically stable features are selected for the experimental design. Under this configuration, the model achieves 93.01 % accuracy, 94 % precision, 93 % recall, and an F1-score of 93 %, maintaining balanced precision and recall. Compared to baseline classifiers that exhibit asymmetric precision–recall behavior, the proposed model produces consistent classification performance on Arabic text. The explanation layer is evaluated using a coherence metric that measures rank agreement between intrinsic LightGBM feature importance and attribution scores. The resulting coherence values (SHAP = 0.726, LIME = 0.3806) indicate that SHAP explanations show higher rank agreement with the model’s internal feature ranking, while LIME captures more localized variation. The findings demonstrate that predictive performance is maintained under feature reduction and that explanation behavior can be evaluated quantitatively.