101. k-Means+++: Outliers-Resistant Clustering
- Author
-
Dan Feldman, Liat Rozenberg, and Adiel Statman
- Subjects
0209 industrial biotechnology ,lcsh:T55.4-60.8 ,Scale (descriptive set theory) ,02 engineering and technology ,numerical_analysis_optimization ,lcsh:QA75.5-76.95 ,Theoretical Computer Science ,Combinatorics ,Reduction (complexity) ,020901 industrial engineering & automation ,Approximation error ,0202 electrical engineering, electronic engineering, information engineering ,lcsh:Industrial engineering. Management engineering ,Cluster analysis ,approximation ,Mathematics ,Numerical Analysis ,outliers ,k-means clustering ,algebra_number_theory ,Computational Mathematics ,Metric space ,Computational Theory and Mathematics ,Outlier ,020201 artificial intelligence & image processing ,lcsh:Electronic computers. Computer science ,Constant (mathematics) ,clustering - Abstract
The k-means problem is to compute a set of k centers (points) that minimizes the sum of squared distances to a given set of n points in a metric space. Arguably, the most common algorithm to solve it is k-means++ which is easy to implement and provides a provably small approximation error in time that is linear in n. We generalize k-means++ to support outliers in two sense (simultaneously): (i) nonmetric spaces, e.g., M-estimators, where the distance dist(p,x) between a point p and a center x is replaced by mindist(p,x),c for an appropriate constant c that may depend on the scale of the input. (ii) k-means clustering with m&ge, 1 outliers, i.e., where the m farthest points from any given k centers are excluded from the total sum of distances. This is by using a simple reduction to the (k+m)-means clustering (with no outliers).
- Published
- 2020