4.6 Article

TWilBert: Pre-trained deep bidirectional transformers for Spanish Twitter

期刊

NEUROCOMPUTING
卷 426, 期 -, 页码 58-69

出版社

ELSEVIER
DOI: 10.1016/j.neucom.2020.09.078

关键词

Contextualized Embedd ngs; Spanish; Twitter; TWilBERT

资金

  1. Spanish Ministerio de Ciencia, Innovacion y Universidades
  2. FEDER [TIN2017-85854-C4-2-R]
  3. Generalitat Valenciana under GiSPRO [PROMETEU/2018/176]
  4. Generalitat Valenciana under GUAITA [INNVA1/2020/61]
  5. Universitat Politecnica de Valencia [PAID-01-17]

向作者/读者索取更多资源

The article introduces TWiLBERT, a specialized variant of BERT for the Spanish language and Twitter domain. By introducing a new Reply Order Prediction signal, the performance in reasoning on Twitter conversation sequences is improved. Across 14 different text classification tasks, TWilBERT outperforms existing systems and Multilingual BERT.
In recent years, the Natural Language Processing community have been moving from uncontextualized word embeddings towards contextualized word embeddings. Among these contextualized architectures, BERT stands out due to its capacity to compute bidirectional contextualized word representations. However, its competitive performance in English downstream tasks is not obtained by its multilingual version when it is applied to other languages and domains. This is especially true in the case of the Spanish language used in Twitter. In this work, we propose TWiLBERT, a specialization of BERT architecture both for the Spanish language and the Twitter domain. Furthermore, we propose a Reply Order Prediction signal to learn inter-sentence coherence in Twitter conversations, which improves the performance of TWilBERT in text classification tasks that require reasoning on sequences of tweets. We perform an extensive evaluation of TWilBERT models on 14 different text classification tasks, such as irony detection, sentiment analysis, or emotion detection. The results obtained by TWilBERT outperform the state-of-the-art systems and Multilingual BERT. In addition, we carry out a thorough analysis of the TWilBERT models to study the reasons of their competitive behavior. We release the pre-trained TWilBERT models used in this paper, along with a framework for training, evaluating, and fine-tuning TWilBERT models. (C) 2020 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据