4.5 Article

ParsBERT: Transformer-based Model for Persian Language Understanding

Journal

NEURAL PROCESSING LETTERS
Volume 53, Issue 6, Pages 3831-3847

Publisher

SPRINGER
DOI: 10.1007/s11063-021-10528-4

Keywords

Persian; Transformers; BERT; Language Models; NLP; NLU

Ask authors/readers for more resources

The rise of pre-trained language models has ushered in a new era in the field of Natural Language Processing, allowing for the development of powerful language models. This paper introduces a monolingual BERT model for the Persian language (ParsBERT) along with a large dataset for various NLP tasks, achieving state-of-the-art performance.
The surge of pre-trained language models has begun a new era in the field of Natural Language Processing (NLP) by allowing us to build powerful language models. Among these models, Transformer-based models such as BERT have become increasingly popular due to their state-of-the-art performance. However, these models are usually focused on English, leaving other languages to multilingual models with limited resources. This paper proposes a monolingual BERT for the Persian language (ParsBERT), which shows its state-of-the-art performance compared to other architectures and multilingual models. Also, since the amount of data available for NLP tasks in Persian is very restricted, a massive dataset for different NLP tasks as well as pre-training the model is composed. ParsBERT obtains higher scores in all datasets, including existing ones and gathered ones, and improves the state-of-the-art performance by outperforming both multilingual BERT and other prior works in Sentiment Analysis, Text Classification, and Named Entity Recognition tasks.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available