Fast k-nearest neighbor classification using cluster-based trees

@article{Zhang2004FastKN,
  title={Fast k-nearest neighbor classification using cluster-based trees},
  author={Bin Zhang and S. Srihari},
  journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
  year={2004},
  volume={26},
  pages={525-528}
}
  • Bin Zhang, S. Srihari
  • Published 2004
  • Computer Science, Medicine
  • IEEE Transactions on Pattern Analysis and Machine Intelligence
Most fast k-nearest neighbor (k-NN) algorithms exploit metric properties of distance measures for reducing computation cost and a few can work effectively on both metric and nonmetric measures. [...] Key Method A mechanism of early decision making and minimal side-operations for choosing searching paths largely contribute to the efficiency of the algorithm. The algorithm is evaluated through extensive experiments over standard NIST and MNIST databases.Expand
A K-Nearest Neighbor Algorithm based on cluster in text classification
TLDR
Clustering was utilized in training set through an improved K-mean approach to select the most representative samples as cluster center and the comparability between the testing samples and the central vector of each cluster was presented. Expand
A Fast Hybrid k-NN Classifier Based on Homogeneous Clusters
TLDR
Experimental results show that the proposed method is fast and accurate, and, in addition, has low pre-processing computational cost. Expand
Efficient $$k$$k-NN classification based on homogeneous clusters
TLDR
The paper proposes an efficient model for fast and accurate nearest neighbor classification that consists of a non-parametric cluster-based preprocessing algorithm that constructs a two-level speed-up data structure and algorithms that access this structure to perform the classification. Expand
Fast Nearest Neighbor Classification using Class-Based Clustering
TLDR
Experimental results show that the proposed parameter-free classification algorithm is very easy to operate and gives consideration to speed and accuracy. Expand
The Research on An Adaptive K-Nearest Neighbors Classifier
k-Nearest neighbor (KNNC) classifier is the most popular non-parametric classifier. But it requires much classification time to search k nearest neighbors of an unlabelled object point, which badlyExpand
The Research on an Adaptive k-Nearest Neighbors Classifier
  • Xiaopeng Yu, Xiaogao Yu
  • Mathematics, Computer Science
  • 2006 5th IEEE International Conference on Cognitive Informatics
  • 2006
TLDR
An adaptive k-nearest neighbors classifier (AKNNC) is proposed that can find k nearest neighbors of the unlabelled point in a small hypersphere in order to improve the efficiencies and classify the point. Expand
Fast k most similar neighbor classifier for mixed data (tree k-MSN)
TLDR
An approximate fast k most similar neighbor classifier, for mixed data and similarity functions that do not satisfy metric properties, based on a tree structure (Tree k-MSN) is proposed. Expand
A Simple Noise-Tolerant Abstraction Algorithm for Fast k-NN Classification
TLDR
This paper shows that the k-means clustering algorithm can be used as a noise-tolerant Data Reduction Technique and shows that if the reduced dataset includes the k -means centroids as representatives of the initial data, performance is not negatively affected as much by the addition of noise. Expand
An Adaptive Hybrid and Cluster-Based Model for Speeding Up the k-NN Classifier
TLDR
A cluster-based classification model for speeding up the k-NN classifier that consists of a simple data structure and a hybrid, adaptive algorithm that accesses this structure and attempts to classify an incoming item using the nearest cluster centroids. Expand
Classification Based on Attribute Positive Correlation and Average Similarity of Nearest Neighbors
The K-Nearest Neighbor algorithm (KNN) is a method for classifying objects based on the k closest training objects. An object is classified by a majority vote of its nearest neighbors. "Closeness" isExpand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 29 REFERENCES
A Fast k Nearest Neighbor Finding Algorithm Based on the Ordered Partition
  • Baek S. Kim, Song B. Park
  • Mathematics, Computer Science
  • IEEE Transactions on Pattern Analysis and Machine Intelligence
  • 1986
TLDR
It is proved that the proposed algorithm can find k nearest neighbors in a constant expected time and is distribution free, and only 4.6 distance calculations were required to find a nearest neighbor among 10 000 samples drawn from a bivariate normal distribution. Expand
Fast Nearest-Neighbor Search in Dissimilarity Spaces
TLDR
An exact probabilistic analysis of this family of algorithms that proves its O(1) asymptotic average complexity measured in the number of dissimilarity calculations is presented. Expand
A Branch and Bound Algorithm for Computing k-Nearest Neighbors
TLDR
The method of branch and bound is implemented in the present algorithm to facilitate rapid calculation of the k-nearest neighbors, by eliminating the necesssity of calculating many distances. Expand
Strategies for efficient incremental nearest neighbor search
TLDR
It is shown that incremental search can be implemented as a sequence of invocations of a previously published non-incremental algorithm, and a new incremental search algorithm is presented which finds the next nearest neighbor more efficiently by eliminating redundant computations. Expand
An algorithm for a selective nearest neighbor decision rule (Corresp.)
A procedure is introduced to approximate nearest neighbor (INN) decision boundaries. The algorithm produces a selective subset of the original data so that 1) the subset is consistent, 2) theExpand
A simple algorithm for nearest neighbor search in high dimensions
The problem of finding the closest point in high-dimensional spaces is common in pattern recognition. Unfortunately, the complexity of most existing search algorithms, such as k-d tree and R-tree,Expand
Finding Prototypes For Nearest Neighbor Classifiers
TLDR
An algorithm is given to find prototypes for a nearest neighbor classifier, which is to start with every sample in a training set as a prototype, and then successively merge any two nearest prototypes of the same class so long as the recognition rate is not downgraded. Expand
Discriminant Adaptive Nearest Neighbor Classification
TLDR
A locally adaptive form of nearest neighbor classification is proposed to try to finesse this curse of dimensionality, and a method for global dimension reduction is proposed, that combines local dimension information. Expand
An algorithm for finding nearest neighbours in (approximately) constant average time
Abstract A new algorithm is proposed which finds the Nearest Neighbour of a given sample in approximately constant average time complexity (i.e. independent of the data set size). The algorithm doesExpand
The Condensed Nearest Neighbor Rule
Since, by (8) pertaining to the nearest neighbor decision rule (NN rule). We briefly review the NN rule and then describe the CNN rule. The NN rule['l-[ " I assigns an unclassified sample to the sameExpand
...
1
2
3
...