A naïve implementation of k-means clustering requires computing for each of the n data points the distance to each of the k cluster centers, which can result in fairly slow execution. However, by storing distance information obtained by earlier computations as well as information about distances between cluster centers, the triangle inequality can be exploited in different ways to reduce the number of needed distance computations, e.g. [3, 4, 5, 7, 11]. In this paper I present an improvement of the Exponion method [11] that generally accelerates the computations. Furthermore, by evaluating several methods on a fairly wide range of artificial data sets, I derive a kind of map, for which data set parameters which method (often) yields the low...
In this paper, we present a novel algorithm for performing k-means clustering. It organizes all the ...
Due to the progressive growth of the amount of data available in a wide variety of scientific fields...
K-means algorithm is very sensitive in initial starting points. Because of initial starting points g...
The -means algorithm is by far the most widely used method for discovering clusters in data. We show...
The popular k-means algorithm is used to discover clusters in vector data automatically. We present ...
We present new algorithms for the k-means clustering problem. They use the kd-tree data structure to...
We present new algorithms for the k-means clustering problem. They use the kd-tree data structure to...
Cluster analysis method is one of the most analytical methods of data mining. The method will direct...
One of the most frequent ways how to cluster data is k-means. The standard way of solving the proble...
K-means clustering is a very popular clustering technique, which is used in numerous applications. ...
Probably the most famous clustering formulation is k-means. This is the focus today. Note: k-means i...
Abstract—This paper introduces an optimized version of the standard K-Means algorithm. The optimizat...
AbstractÐIn k-means clustering, we are given a set of n data points in d-dimensional space Rd and an...
We study the problem of finding an optimum clustering, a problem known to be NP-hard. Existing liter...
We present polynomial upper and lower bounds on the number of iterations performed by the k-means me...
In this paper, we present a novel algorithm for performing k-means clustering. It organizes all the ...
Due to the progressive growth of the amount of data available in a wide variety of scientific fields...
K-means algorithm is very sensitive in initial starting points. Because of initial starting points g...
The -means algorithm is by far the most widely used method for discovering clusters in data. We show...
The popular k-means algorithm is used to discover clusters in vector data automatically. We present ...
We present new algorithms for the k-means clustering problem. They use the kd-tree data structure to...
We present new algorithms for the k-means clustering problem. They use the kd-tree data structure to...
Cluster analysis method is one of the most analytical methods of data mining. The method will direct...
One of the most frequent ways how to cluster data is k-means. The standard way of solving the proble...
K-means clustering is a very popular clustering technique, which is used in numerous applications. ...
Probably the most famous clustering formulation is k-means. This is the focus today. Note: k-means i...
Abstract—This paper introduces an optimized version of the standard K-Means algorithm. The optimizat...
AbstractÐIn k-means clustering, we are given a set of n data points in d-dimensional space Rd and an...
We study the problem of finding an optimum clustering, a problem known to be NP-hard. Existing liter...
We present polynomial upper and lower bounds on the number of iterations performed by the k-means me...
In this paper, we present a novel algorithm for performing k-means clustering. It organizes all the ...
Due to the progressive growth of the amount of data available in a wide variety of scientific fields...
K-means algorithm is very sensitive in initial starting points. Because of initial starting points g...