TY - JOUR
T1 - Coresets for kernel clustering
AU - Jiang, Shaofeng H.C.
AU - Krauthgamer, Robert
AU - Lou, Jianing
AU - Zhang, Yubo
N1 - Publisher Copyright: © The Author(s), under exclusive licence to Springer Science+Business Media LLC, part of Springer Nature 2024.
PY - 2024/4/22
Y1 - 2024/4/22
N2 - We devise coresets for kernel k-Means with a general kernel, and use them to obtain new, more efficient, algorithms. Kernel k-Means has superior clustering capability compared to classical k-Means, particularly when clusters are non-linearly separable, but it also introduces significant computational challenges. We address this computational issue by constructing a coreset, which is a reduced dataset that accurately preserves the clustering costs. Our main result is a coreset for kernel k-Means that works for a general kernel and has size poly(kϵ-1). Our new coreset both generalizes and greatly improves all previous results; moreover, it can be constructed in time near-linear in n. This result immediately implies new algorithms for kernel k-Means, such as a (1+ϵ)-approximation in time near-linear in n, and a streaming algorithm using space and update time poly(kϵ-1logn). We validate our coreset on various datasets with different kernels. Our coreset performs consistently well, achieving small errors while using very few points. We show that our coresets can speed up kernel K-MEANS++ (the kernelized version of the widely used K-MEANS++ algorithm), and we further use this faster kernel K-MEANS++ for spectral clustering. In both applications, we achieve significant speedup and a better asymptotic growth while the error is comparable to baselines that do not use coresets.
AB - We devise coresets for kernel k-Means with a general kernel, and use them to obtain new, more efficient, algorithms. Kernel k-Means has superior clustering capability compared to classical k-Means, particularly when clusters are non-linearly separable, but it also introduces significant computational challenges. We address this computational issue by constructing a coreset, which is a reduced dataset that accurately preserves the clustering costs. Our main result is a coreset for kernel k-Means that works for a general kernel and has size poly(kϵ-1). Our new coreset both generalizes and greatly improves all previous results; moreover, it can be constructed in time near-linear in n. This result immediately implies new algorithms for kernel k-Means, such as a (1+ϵ)-approximation in time near-linear in n, and a streaming algorithm using space and update time poly(kϵ-1logn). We validate our coreset on various datasets with different kernels. Our coreset performs consistently well, achieving small errors while using very few points. We show that our coresets can speed up kernel K-MEANS++ (the kernelized version of the widely used K-MEANS++ algorithm), and we further use this faster kernel K-MEANS++ for spectral clustering. In both applications, we achieve significant speedup and a better asymptotic growth while the error is comparable to baselines that do not use coresets.
UR - http://www.scopus.com/inward/record.url?scp=85191099214&partnerID=8YFLogxK
U2 - 10.1007/s10994-024-06540-z
DO - 10.1007/s10994-024-06540-z
M3 - مقالة
SN - 0885-6125
JO - Machine Learning
JF - Machine Learning
ER -