Relevance-Redundancy Dominance: a Threshold-Free Approach to Filter-Based Feature Selection
Publication Type:
Refereed Conference Meeting Proceeding
Abstract:
Feature selection is used to select a subset of relevant features
in machine learning, and is vital for simplication, improving eciency
and reducing overtting. In lter-based feature selection, a statistic such
as correlation or entropy is computed between each feature and the target
variable to evaluate feature relevance. A relevance threshold is typically
used to limit the set of selected features, and features can also be removed
based on redundancy (similarity to other features). Some methods are
designed for use with a specic statistic or certain types of data. We
present a new lter-based method called Relevance-Redundancy Domi-
nance that applies to mixed data types, can use a wide variety of statis-
tics, and does not require a threshold. In experiments it outperformed
published methods on credit data.
Conference Name:
24th Irish Conference on Artificial Intelligence and Cognitive Science (AICS) 2016
Proceedings:
to be appear
Digital Object Identifer (DOI):
to be appear
Publication Date:
21/09/2016
Conference Location:
Ireland
Research Group:
Institution:
National University of Ireland, Cork (UCC)
Open access repository:
No