Wrappers for Feature Subset Selection

Abstract

In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set, a feature subset selection method should consider how the algorithm and the training set interact. We explore the relation between optimal feature subset selection and relevance. Our wrapper method searches for an optimal feature subset tailored to a particular algorithm and a domain. We study the strengths and weaknesses of the wrapper approach and show a series of improved designs. We compare the wrapper approach to induction without feature subset selection and to Relief, a filter approach to feature subset selection. Significant improvement in accuracy is achieved for some datasets for the two families of induction algorithms used: decision trees and Naive-Bayes. @ 1997 Elsevier Science B.V.

DOI: 10.1016/S0004-3702(97)00043-X

Extracted Key Phrases

35 Figures and Tables

0200400600'96'98'00'02'04'06'08'10'12'14'16'18
Citations per Year

8,135 Citations

Semantic Scholar estimates that this publication has 8,135 citations based on the available data.

See our FAQ for additional information.

Cite this paper

@article{Kohavi1997WrappersFF, title={Wrappers for Feature Subset Selection}, author={Ron Kohavi and George H. John}, journal={Artif. Intell.}, year={1997}, volume={97}, pages={273-324} }