Journal
APPLIED SCIENCES-BASEL
Volume 12, Issue 17, Pages -Publisher
MDPI
DOI: 10.3390/app12178398
Keywords
text augmentation; transfer learning; Bidirectional Encoder Representation of Transformers; fake news classification
Ask authors/readers for more resources
Researchers proposed a method using text augmentation technique and BERT language model to generate an augmented dataset, which addressed the issue of bias and improved fake news detection performance.
Fake news detection techniques are a topic of interest due to the vast abundance of fake news data accessible via social media. The present fake news detection system performs satisfactorily on well-balanced data. However, when the dataset is biased, these models perform poorly. Additionally, manual labeling of fake news data is time-consuming, though we have enough fake news traversing the internet. Thus, we introduce a text augmentation technique with a Bidirectional Encoder Representation of Transformers (BERT) language model to generate an augmented dataset composed of synthetic fake data. The proposed approach overcomes the issue of minority class and performs the classification with the AugFake-BERT model (trained with an augmented dataset). The proposed strategy is evaluated with twelve different state-of-the-art models. The proposed model outperforms the existing models with an accuracy of 92.45%. Moreover, accuracy, precision, recall, and f1-score performance metrics are utilized to evaluate the proposed strategy and demonstrate that a balanced dataset significantly affects classification performance.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available