#### Filter Results:

- Full text PDF available (28)

#### Publication Year

1993

2015

- This year (0)
- Last 5 years (1)
- Last 10 years (5)

#### Publication Type

#### Co-author

#### Journals and Conferences

#### Data Set Used

#### Key Phrases

Learn More

- Ron Kohavi, George H. John
- Artif. Intell.
- 1997

In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set, a feature subset selection method should consider how the… (More)

- George H. John, Pat Langley
- UAI
- 1995

When modeling a probability distribution with a Bayesian network, we are faced with the problem of how to handle continuous vari ables. Most previous work has either solved the problem by discretizing, or assumed that the data are generated by a single Gaussian. In this paper we abandon the normality as sumption and instead use statistical methods for… (More)

- George H. John, Ron Kohavi, Karl Pfleger
- ICML
- 1994

We address the problem of nding a subset of features that allows a supervised induc tion algorithm to induce small high accuracy concepts We examine notions of relevance and irrelevance and show that the de nitions used in the machine learning literature do not adequately partition the features into useful categories of relevance We present de ni tions for… (More)

- George H. John, Pat Langley
- KDD
- 1996

As data warehouses grow to the point where one hundred gigabytes is considered small, the computational efficiency of data-mining algorithms on large databases becomes increasingly important. Using a sample from the database can speed up the datamining process, but this is only acceptable if it does not reduce the quality of the mined knowledge. To this… (More)

- George H. John
- KDD
- 1995

Finding and removing outliers is an important problem in data mining. Errors in large databases can be extremely common, so an important property of a data mining algorithm is robustness with respect to errors in the database. Most sophisticated methods in machine learning address this problem to some extent, but not fully, and can be improved by addressing… (More)

- Ron Kohavi, George H. John, Richard Long, David Manley, Karl Pfleger
- ICTAI
- 1994

We present MLC ++ , a library of C ++ classes and tools for supervised Machine Learning. While MLC ++ provides general learning algorithms that can be used by end users, the main objective is to provide researchers and experts with a wide variety of tools that can accelerate algorithm development, increase software reliability, provide comparison tools, and… (More)

- George H. John
- 1995

When modeling a probability distribution with a Bayesian network, we are faced with the problem of how to handle continuous variables. Most previous work has either solved the problem by discretizing, or assumed that the data are generated by a single Gaussian. In this paper we abandon the normality assumption and instead use statistical methods for… (More)

- George H. John
- AAAI
- 1994

The most popular delayed reinforcement learning technique, Q-learning (Watkins 1989)) estimates the future reward expected from executing each action in every state. If these estimates are correct, then an agent can use them to select the action with maximal expected future reward in each state, and thus perform optimally. Watkins has proved that Q-learning… (More)

- George H. John, Brian Lent
- KDD
- 1997

When mining large databases, the data extraction problem and the interface between the database and data mining algorithm become important issues. Rather than giving a mining algorithm full access to a database (by extracting to a flat file or other directlyaccessible data structure), we propose the SQL Interface Protocol (SIP), which is a framework for… (More)

- George H. John
- AISTATS
- 1995

We present a new method for the induction of classiication trees with linear dis-criminants as the partitioning function at each internal node. This paper presents two main contributions: rst, a novel objective function called soft entropy which is used to identify optimal coeecients for the linear discriminants, and second, a novel method for removing… (More)