- Published 1997 in Artif. Intell.

In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set, a feature subset selection method should consider how the algorithm and the training set interact. We explore the relation between optimal feature subset selection and relevance. Our wrapper method searches for an optimal feature subset tailored to a particular algorithm and a domain. We study the strengths and weaknesses of the wrapper approach and show a series of improved designs. We compare the wrapper approach to induction without feature subset selection and to Relief, a filter approach to feature subset selection. Significant improvement in accuracy is achieved for some datasets for the two families of induction algorithms used: decision trees and Naive-Bayes. @ 1997 Elsevier Science B.V.

Citations per Year

Semantic Scholar estimates that this publication has **8,135** citations based on the available data.

See our **FAQ** for additional information.

Showing 1-10 of 3,021 extracted citations

Highly Influenced

6 Excerpts

Highly Influenced

10 Excerpts

Highly Influenced

10 Excerpts

Highly Influenced

11 Excerpts

Highly Influenced

20 Excerpts

Highly Influenced

20 Excerpts

Highly Influenced

9 Excerpts

Highly Influenced

5 Excerpts

Highly Influenced

4 Excerpts

Highly Influenced

12 Excerpts

@article{Kohavi1997WrappersFF,
title={Wrappers for Feature Subset Selection},
author={Ron Kohavi and George H. John},
journal={Artif. Intell.},
year={1997},
volume={97},
pages={273-324}
}