4.3 Article

Empirical Evaluation of Shallow and Deep Learning Classifiers for Arabic Sentiment Analysis

Publisher

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3466171

Keywords

Deep learning; shallow learning; learning curve; embedding; misclassification

Ask authors/readers for more resources

This study presents a detailed comparison of the performance of deep learning and shallow learning models in sentiment analysis of Arabic reviews, and found that deep learning outperforms shallow learning in this task. Additionally, the study evaluates the performance of state-of-the-art models and finds that the Transformer model augmented with araBERT performs the best.
This work presents a detailed comparison of the performance of deep learning models such as convolutional neural networks, long short-term memory, gated recurrent units, their hybrids, and a selection of shallow learning classifiers for sentiment analysis of Arabic reviews. Additionally, the comparison includes state-of-the-art models such as the transformer architecture and the araBERT pre-trained model. The datasets used in this study are multi-dialect Arabic hotel and book review datasets, which are some of the largest publicly available datasets for Arabic reviews. Results showed deep learning outperforming shallow learning for binary and multi-label classification, in contrast with the results of similar work reported in the literature. This discrepancy in outcome was caused by dataset size as we found it to be proportional to the performance of deep learning models. The performance of deep and shallow learning techniques was analyzed in terms of accuracy and F1 score. The best performing shallow learning technique was Random Forest followed by Decision Tree, and AdaBoost. The deep learning models performed similarly using a default embedding layer, while the transformer model performed best when augmented with araBERT.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.3
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available