作者
B Amarnath, S Balamurugan, Appavu Alias
发表日期
2016/11/1
期刊
Journal of Engineering Science and Technology
卷号
11
期号
11
页码范围
1639-1646
简介
Feature selection goal is to get rid of redundant and irrelevant features. The problem of feature subset selection is that of finding a subset of the original features of a dataset, such that an induction algorithm run on data containing only selected features makes a classifier to generate with the highest possible accuracy. High dimensional data can contain a high degree of irrelevant and redundant features which may greatly degrade the performance of learning algorithms. The performance of different feature selectors such as CFS, Chi-Square, Information Gain, Gain Ratio, One R and Symmetrical Uncertainty were evaluated on two different popular classification algorithms namely Decision Tree and Naive Bayesian method. A significant improvement in the performance of DT and NB classifier was shown after reducing the number of both irrelevant and redundant features by the use of different feature ranking methods.
引用总数
201820192020202120222023202442251137