Classification and Regression Trees

  title={Classification and Regression Trees},
  author={L. Breiman and Jerome H. Friedman and Richard A. Olshen and C. J. Stone},
Background. Introduction to Tree Classification. Right Sized Trees and Honest Estimates. Splitting Rules. Strengthening and Interpreting. Medical Diagnosis and Prognosis. Mass Spectra Classification. Regression Trees. Bayes Rules and Partitions. Optimal Pruning. Construction of Trees from a Learning Sample. Consistency. Bibliography. Notation Index. Subject Index. 

Figures from this paper

Classification and Regression Tree Methods

This article discusses the C4.5, CART, CRUISE, GUIDE, and QUEST methods in terms of their algorithms, features, properties, and performances.

Classification and Regression Trees

CART is a method that provides mechanisms for building a custom-specific, nonparametric estimation model based solely on the analysis of measurement project data, called training data.

Classification and Regression Trees

We will call an estimator for the regression function defined by the CART methodology a regression tree. The word CART means classification and regression tree. This chapter will focus only on the

Cost-Sensitive Pruning of Decision Trees

This paper shows how the misclassification costs, a related criterion applied if errors vary in their costs, can be integrated in several well-known pruning techniques.

Survival Trees by Goodness of Split

Abstract A tree-based method for censored survival data is developed, based on maximizing the difference in survival between groups of patients represented by nodes in a binary tree. The method

Selecting the best categorical split for classification trees

Based on a family of splitting criteria for classification trees, methods of selecting the best categorical splits are studied. They are shown to be very useful in reducing the computational

Data Mining Classification : Basic Concepts , Decision Trees , and Model Evaluation

Classification, which is the task of assigning objects to one of several predefined categories, is a pervasive problem that encompasses many diverse applications. Examples include detecting spam

Randomization in Aggregated Classification Trees

This paper discusses and compares different methods for model aggregation, and addresses the problem of finding minimal number of trees sufficient for the forest.

Using Model Trees for Classification

Surprisingly, using this simple transformation the model tree inducer M5′, based on Quinlan's M5, generates more accurate classifiers than the state-of-the-art decision tree learner C5.0, particularly when most of the attributes are numeric.

Classification Based on Tree-Structured Allocation Rules

The authors consider the problem of classifying an unknown observation into 1 of several populations by using tree-structured allocation rules. Although many parametric classification procedures are



Induction over large data bases

Techniques for discovering rules by induction from large collections of instances are developed. These are based on an iterative scheme for dividing the instances into two sets, only one of which

Application of information theory to the construction of efficient decision trees

This heuristic approach to the problem of conversion of decision tables to decision trees is treated and has low design complexity and yet provides near-optimal decision trees.

A Partitioning Algorithm with Application in Pattern Classification and the Optimization of Decision Trees

A two-stage algorithm that obtains a sufficient partition suboptimally, either by methods suggested in the paper or developed elsewhere, and optimizes the results of the first stage through a dynamic programming approach is proposed.

Hierarchical Classifier Design Using Mutual Information

A nonparametric algorithm is presented for the hierarchical partitioning of the feature space that generates an efficient partitioning tree for specified probability of error by maximizing the amount of average mutual information gain at each partitioning step.

Pattern classification and scene analysis

The topics treated include Bayesian decision theory, supervised and unsupervised learning, nonparametric techniques, discriminant analysis, clustering, preprosessing of pictorial data, spatial filtering, shape description techniques, perspective transformations, projective invariants, linguistic procedures, and artificial intelligence techniques for scene analysis.

An Algorithm for Constructing Optimal Binary Decision Trees

It is shown that an optimal tree can be recursively constructed through the application of invariant imbedding (dynamic programming) and an algorithm is detailed which embodies this recursive approach.

Regression and ANOVA with Zero-One Data: Measures of Residual Variation

Abstract We consider regression situations for which the response variable is dichotomous. The most common analysis fits successively richer linear logistic models and measures the residual variation

Some methods for classification and analysis of multivariate observations

The main purpose of this paper is to describe a process for partitioning an N-dimensional population into k sets on the basis of a sample. The process, which is called 'k-means,' appears to give

Constructing Optimal Binary Decision Trees is NP-Complete