Sentiment Analysis for Arabic Using Deep Learning
Loading...

Date
2026
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Springer Science and Business Media Deutschland GmbH
Open Access Color
Green Open Access
No
OpenAIRE Downloads
OpenAIRE Views
Publicly Funded
No
Abstract
With the explosive growth of digital communication, understanding sentiment in online content has become increasingly critical for a wide range of applications, from customer feedback analysis to social media monitoring. However, sentiment analysis for Arabic presents unique challenges due to the language's rich morphology, diverse dialects, and complex syntactic structures. These challenges are further amplified in multimodal settings, where the fusion of textual, visual, and auditory cues is required to capture the full spectrum of human emotion. To address these issues, this paper introduces a new framework for Arabic Multimodal Sentiment Analysis (AMSA), combining multi-level deep learning approaches across text, audio, and visual modalities. Our approach utilizes state-of-the-art transformer-based architecturees, including Multimodal Transformer (MulT) and Early Fusion models, to tackle both linguistic complexity and multimodal alignment. Specifically, we leverage DeBERTa for extracting rich textual features, ViT (Vision Transformer) for visual cues, and Whisper for capturing nuanced audio signals, creating robust and contextualized representations. Experimental results on a curated Arabic multimodal dataset demonstrate the effectiveness of this approach, with our proposed MulT model achieving an F1 score of 72.73%, reflecting a substantial improvement of 13.98% in F1 score and 14.6% in accuracy over existing baselines. These findings highlight the power of cross-modal attention mechanisms and early fusion strategies in accurately capturing subtle sentiments across multiple modalities. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2026.
Description
Keywords
Arabic Sentiment Analysis, Deep Learning, Early Fusion, Multimodal Sentiment Analysis, Multimodal Transformer
Fields of Science
Citation
WoS Q
N/A
Scopus Q
Q4

OpenCitations Citation Count
N/A
Source
Lecture Notes in Networks and Systems
Volume
1594 LNNS
Issue
Start Page
65
End Page
73
Collections
PlumX Metrics
Citations
Scopus : 0
Page Views
5
checked on Feb 25, 2026
Google Scholar™


