A novel LightGBM model for Arabic spam detection integrated with XAI for enhanced explainability
Arabic spam detection presents technical challenges due to linguistic variability, feature sparsity, and the limited availability of interpretable classification systems. Several machine learning models lack intrinsic interpretability, which reduces transparency in decision-making. This study propos...
Saved in:
| Main Authors: | , , , |
|---|---|
| Format: | Article |
| Language: | en |
| Published: |
Elsevier
2026
|
| Subjects: | |
| Online Access: | http://psasir.upm.edu.my/id/eprint/123346/1/123346.pdf http://psasir.upm.edu.my/id/eprint/123346/ https://www.sciencedirect.com/science/article/pii/S0045790626001047 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Summary: | Arabic spam detection presents technical challenges due to linguistic variability, feature sparsity, and the limited availability of interpretable classification systems. Several machine learning models lack intrinsic interpretability, which reduces transparency in decision-making. This study proposes a spam detection pipeline that combines LightGBM with an integrated explainability layer. Interpretability is incorporated directly into the evaluation process by embedding SHAP and LIME to quantify explanation behavior. The system targets characteristics of Arabic text data by operating in a compressed feature space. From an initial 100-dimensional TF-IDF representation, the top 20 statistically stable features are selected for the experimental design. Under this configuration, the model achieves 93.01 % accuracy, 94 % precision, 93 % recall, and an F1-score of 93 %, maintaining balanced precision and recall. Compared to baseline classifiers that exhibit asymmetric precision–recall behavior, the proposed model produces consistent classification performance on Arabic text. The explanation layer is evaluated using a coherence metric that measures rank agreement between intrinsic LightGBM feature importance and attribution scores. The resulting coherence values (SHAP = 0.726, LIME = 0.3806) indicate that SHAP explanations show higher rank agreement with the model’s internal feature ranking, while LIME captures more localized variation. The findings demonstrate that predictive performance is maintained under feature reduction and that explanation behavior can be evaluated quantitatively. |
|---|
