4.6 Article

Improving the Naive Bayes Classifier via a Quick Variable Selection Method Using Maximum of Entropy

期刊

ENTROPY
卷 19, 期 6, 页码 -

出版社

MDPI
DOI: 10.3390/e19060247

关键词

variable selection; classification; Naive Bayes; imprecise probabilities; uncertainty measures

资金

  1. Spanish Ministerio de Economia y Competitividad
  2. Fondo Europeo de Desarrollo Regional (FEDER) [TEC2015-69496-R]

向作者/读者索取更多资源

Variable selection methods play an important role in the field of attribute mining. The Naive Bayes (NB) classifier is a very simple and popular classification method that yields good results in a short processing time. Hence, it is a very appropriate classifier for very large datasets. The method has a high dependence on the relationships between the variables. The Info-Gain (IG) measure, which is based on general entropy, can be used as a quick variable selection method. This measure ranks the importance of the attribute variables on a variable under study via the information obtained from a dataset. The main drawback is that it is always non-negative and it requires setting the information threshold to select the set of most important variables for each dataset. We introduce here a new quick variable selection method that generalizes the method based on the Info-Gain measure. It uses imprecise probabilities and the maximum entropy measure to select the most informative variables without setting a threshold. This new variable selection method, combined with the Naive Bayes classifier, improves the original method and provides a valuable tool for handling datasets with a very large number of features and a huge amount of data, where more complex methods are not computationally feasible.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

Article Computer Science, Artificial Intelligence

Analysis of Credal-C4.5 for classification in noisy domains

Carlos J. Mantas, Joaquin Abellan, Javier G. Castellano

EXPERT SYSTEMS WITH APPLICATIONS (2016)

Article Computer Science, Artificial Intelligence

A comparative study on base classifiers in ensemble methods for credit scoring

Joaquin Abelian, Javier G. Castellano

EXPERT SYSTEMS WITH APPLICATIONS (2017)

Article Computer Science, Theory & Methods

Extraction of decision rules via imprecise probabilities

Joaquin Abellan, Griselda Lopez, Laura Garach, Javier G. Castellano

INTERNATIONAL JOURNAL OF GENERAL SYSTEMS (2017)

Article Computer Science, Artificial Intelligence

A Random Forest approach using imprecise probabilities

Joaquin Abellan, Carlos J. Mantas, Javier G. Castellano

KNOWLEDGE-BASED SYSTEMS (2017)

Article Computer Science, Artificial Intelligence

AdaptativeCC4.5: Credal C4.5 with a rough class noise estimator

Joaquin Abelian, Carlos J. Mantas, Javier G. Castellano

EXPERT SYSTEMS WITH APPLICATIONS (2018)

Article Computer Science, Artificial Intelligence

Increasing diversity in random forest learning algorithm via imprecise probabilities

Joaquin Abellan, Carlos J. Mantas, Javier G. Castellano, SerafIn Moral-Garcia

EXPERT SYSTEMS WITH APPLICATIONS (2018)

Article Computer Science, Artificial Intelligence

Using Credal-C4.5 with Binary Relevance for Multi-Label Classification

Serafin Moral-Garcia, Carlos J. Mantas, Javier G. Castellano, Joaquin Abellan

JOURNAL OF INTELLIGENT & FUZZY SYSTEMS (2018)

Article Physics, Multidisciplinary

Decision Tree Ensemble Method for Analyzing Traffic Accidents of Novice Drivers in Urban Areas

Serafin Moral-Garcia, Javier G. Castellano, Carlos J. Mantas, Alfonso Montella, Joaquin Abellan

ENTROPY (2019)

Article Biochemistry & Molecular Biology

Combining gene expression data and prior knowledge for inferring gene regulatory networks via Bayesian networks using structural restrictions

Luis M. de Campos, Andres Cano, Javier G. Castellano, Serafin Moral

STATISTICAL APPLICATIONS IN GENETICS AND MOLECULAR BIOLOGY (2019)

Article Computer Science, Artificial Intelligence

Bagging of credal decision trees for imprecise classification

S. Moral-Garcia, Carlos J. Mantas, Javier G. Castellano, Maria D. Benitez, Joaquin Abellan

EXPERT SYSTEMS WITH APPLICATIONS (2020)

Article Computer Science, Artificial Intelligence

On the Use of m-Probability-Estimation and Imprecise Probabilities in the Naive Bayes Classifier

Javier G. Castellano, Serafin Moral-Garcia, Carlos J. Mantas, Joaquin Abellan

INTERNATIONAL JOURNAL OF UNCERTAINTY FUZZINESS AND KNOWLEDGE-BASED SYSTEMS (2020)

Article Automation & Control Systems

A Decision Support Tool for Credit Domains: Bayesian Network with a Variable Selector Based on Imprecise Probabilities

Javier G. Castellano, Serafin Moral-Garcia, Carlos J. Mantas, Maria D. Benitez, Joaquin Abellan

Summary: A Bayesian Network is a graphical structure with conditional probability tables that allows for calculating probabilities between different features, particularly useful in credit scoring and risk analysis.

INTERNATIONAL JOURNAL OF FUZZY SYSTEMS (2021)

Article Computer Science, Artificial Intelligence

Using extreme prior probabilities on the Naive Credal Classifier

Serafin Moral-Garcia, Javier G. Castellano, Carlos J. Mantas, Joaquin Abellan

Summary: The Naive Credal Classifier (NCC) was the first proposed method for Imprecise Classification. In this work, a new version of NCC called Extreme Prior Naive Credal Classifier (EP-NCC) is proposed, which considers the prior probabilities of the class variable when estimating the conditional probabilities. It is shown that EP-NCC achieves more informative predictions without increasing the risk of errors. Experimental analysis demonstrates that EP-NCC outperforms NCC and achieves statistically equivalent results to the existing algorithm for Imprecise Classification based on decision trees, while being computationally simpler.

KNOWLEDGE-BASED SYSTEMS (2022)

Article Mathematics, Interdisciplinary Applications

A New Robust Classifier on Noise Domains: Bagging of Credal C4.5 Trees

Joaquin Abellan, Javier G. Castellano, Carlos J. Mantas

COMPLEXITY (2017)

暂无数据