• Corpus ID: 211677991

Advanced kNN: A Mature Machine Learning Series

@article{Asim2020AdvancedKA,
  title={Advanced kNN: A Mature Machine Learning Series},
  author={Muhammad Asim and Muaaz Zakria},
  journal={ArXiv},
  year={2020},
  volume={abs/2003.00415}
}
k-nearest neighbour (kNN) is one of the most prominent, simple and basic algorithm used in machine learning and data mining. However, kNN has limited prediction ability, i.e., kNN cannot predict any instance correctly if it does not belong to any of the predefined classes in the training data set. The purpose of this paper is to suggest an Advanced kNN (A-kNN) algorithm that will be able to classify an instance as unknown, after verifying that it does not belong to any of the predefined classes… 
Detection of Coronary Heart Disease Using Modified K-NN Method with Recursive Feature Elimination
TLDR
This study used the KNN method by developing the weight of each class to increase accuracy, and the results prove the effectiveness of the Weighted KNN with feature selection.
Hierarchical models of network traffic
The spread of malware is constantly growing, and along with the transformation of the world into digital form, this problem is an increasingly essential and discussed topic. There are various ways to

References

SHOWING 1-10 OF 23 REFERENCES
A New Classification Algorithm Using Mutual Nearest Neighbors
TLDR
A new learning algorithm under the framework of kNN that adopts mutual nearest neighbors, rather than k nearestNeighbors, to determine the class labels of unknown instances, and shows that the classification performance achieved by the proposed method is better than the traditional one.
KNN Model-Based Approach in Classification
TLDR
This paper proposes a novel kNN type method for classification that reduces the dependency on k, makes classification faster, and compares well with C5.0 and kNN in terms of classification accuracy.
Cost-sensitive KNN classification
TLDR
Two efficient cost-sensitive KNN classification models are designed, referred to Direct-CS-KNN classifier and Distance- CS-Knn classifier, which are further improved with extant strategies, such as smoothing, minimum-cost k-value selection, feature selection and ensemble selection.
Learning k for kNN Classification
TLDR
Experimental results showed that the proposed Correlation Matrix kNN (CM-kNN) classification was more accurate and efficient than existing kNN methods in data-mining applications, such as classification, regression, and missing data imputation.
An Improved k-Nearest Neighbor Algorithm for Text Categorization
TLDR
An improved kNN algorithm is proposed, which uses different numbers of nearest neighbors for different categories, rather than a fixed number across all categories, and is promising for some cases, where estimating the parameter k via cross-validation is not allowed.
kNN Algorithm with Data-Driven k Value
TLDR
The experimental results show that the proposed kNN algorithm is better than the state-of-the-art algorithms in terms of different learning tasks, such as classification, regression, and missing value imputation.
A novel kNN algorithm with data-driven k parameter computation
TLDR
An example-driven k-parameter computation that identifies different k values for different test samples in kNN prediction applications, such as classification, regression and missing data imputation is studied.
k-Nearest Neighbour Classifiers
TLDR
An overview of techniques for Nearest Neighbour classification focusing onMechanisms for assessing similarity (distance), computational issues in identifying nearest neighbours and mechanisms for reducing the dimen-sion of the data is presented.
The Condensed Nearest Neighbor Rule
Since, by (8) pertaining to the nearest neighbor decision rule (NN rule). We briefly review the NN rule and then describe the CNN rule. The NN rule['l-[ " I assigns an unclassified sample to the same
Probably correct k-nearest neighbor search in high dimensions
TLDR
A probably correct approach, in which the correct set of k-nearest neighbors is obtained in high probability, is proposed for greatly reducing the searching time, and the marginal distribution of the k th nearest neighbors in low dimensions is exploited from the stored data.
...
1
2
3
...