k-Nearest Neighbors Task
Jump to navigation
Jump to search
A k-Nearest Neighbors Task is a search task that finds neighbors within a distance.
- Context:
- It can be solved by a Nearest Neighbors System (that implements a nearest neighbors algorithm).
- See: Eager Model-based Learning Task.
References
2016
- http://scikit-learn.org/stable/modules/neighbors.html
- QUOTE: Unsupervised nearest neighbors is the foundation of many other learning methods, notably manifold learning and spectral clustering. Supervised neighbors-based learning comes in two flavors: classification for data with discrete labels, and regression for data with continuous labels. The principle behind nearest neighbor methods is to find a predefined number of training samples closest in distance to the new point, and predict the label from these. The number of samples can be a user-defined constant (k-nearest neighbor learning), or vary based on the local density of points (radius-based neighbor learning). The distance can, in general, be any metric measure: standard Euclidean distance is the most common choice. Neighbors-based methods are known as non-generalizing machine learning methods, since they simply “remember” all of its training data (possibly transformed into a fast indexing structure such as a Ball Tree or KD Tree.). Despite its simplicity, nearest neighbors has been successful in a large number of classification and regression problems, including handwritten digits or satellite image scenes. Being a non-parametric method, it is often successful in classification situations where the decision boundary is very irregular.