This paper addresses the issue of numerical computation in machine learning domains where one needs to evaluate similarity metrics. Examples of these domains include dimensionality reduction with kernels, spectral clustering and Gaussian processes. The paper presents a solution strategy based on Krylov subspace iteration and fast N-body learning methods. The main benefit of this strategy is that it is very general and easy to adapt to new domains involving similarity metrics. The experiments show significant gains in computation and storage on datasets arising in image segmentation, object detection and dimensionality reduction. The paper also presents theoretical bounds on the stability of these iterative methods.
Machine learning model training time can be significantly reduced by using dimensionality reduction ...
We describe a recursive algorithm to quickly compute the N nearest neighbors according to a similari...
Recently, many machine learning problems rely on a valuable tool: metric learning. However, in many ...
A typical machine learning algorithm takes advantage of training data to discover patterns among obs...
Similarity and metric learning provides a principled approach to construct a task-specific similarit...
textA large number of machine learning algorithms are critically dependent on the underlying distanc...
Metric learning has been shown to significantly improve the accuracy of k-nearest neighbor (kNN) cla...
textMachine learning techniques are now essential for a diverse set of applications in computer visi...
Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Comput...
When the data vectors are high-dimensional it is computationally infeasible to use data analysis or ...
Many similarity-based clustering methods work in two separate steps including similarity matrix comp...
International audienceSimilarity between objects plays an important role in both human cognitive pro...
This paper presents two metrics for the Nearest Neighbor Classifier that share the property of being...
This work explores statistical properties of machine learning algorithms from different perspectives...
A good measure of similarity between data points is crucial to many tasks in machine learning. Simil...
Machine learning model training time can be significantly reduced by using dimensionality reduction ...
We describe a recursive algorithm to quickly compute the N nearest neighbors according to a similari...
Recently, many machine learning problems rely on a valuable tool: metric learning. However, in many ...
A typical machine learning algorithm takes advantage of training data to discover patterns among obs...
Similarity and metric learning provides a principled approach to construct a task-specific similarit...
textA large number of machine learning algorithms are critically dependent on the underlying distanc...
Metric learning has been shown to significantly improve the accuracy of k-nearest neighbor (kNN) cla...
textMachine learning techniques are now essential for a diverse set of applications in computer visi...
Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Comput...
When the data vectors are high-dimensional it is computationally infeasible to use data analysis or ...
Many similarity-based clustering methods work in two separate steps including similarity matrix comp...
International audienceSimilarity between objects plays an important role in both human cognitive pro...
This paper presents two metrics for the Nearest Neighbor Classifier that share the property of being...
This work explores statistical properties of machine learning algorithms from different perspectives...
A good measure of similarity between data points is crucial to many tasks in machine learning. Simil...
Machine learning model training time can be significantly reduced by using dimensionality reduction ...
We describe a recursive algorithm to quickly compute the N nearest neighbors according to a similari...
Recently, many machine learning problems rely on a valuable tool: metric learning. However, in many ...