TY - JOUR
T1 - Convex clustering: Model, theoretical guarantee and efficient algorithm
AU - Sun, Defeng
AU - Toh, Kim Chuan
AU - Yuan, Yancheng
N1 - Funding Information:
We thank the three referees for the helpful suggestions to improve this paper. Defeng Sun is supported in part by Hong Kong Research Grant Council under grant PolyU 153014/18P. Kim-Chuan Toh is supported by the Academic Research Fund of the Ministry of Education, Singapore, under grant R-146-000-257-112.
Publisher Copyright:
© 2021 Microtome Publishing. All rights reserved.
PY - 2021/1
Y1 - 2021/1
N2 - Clustering is a fundamental problem in unsupervised learning. Popular methods like Kmeans, may suffer from poor performance as they are prone to get stuck in its local minima. Recently, the sum-of-norms (SON) model (also known as the convex clustering model) has been proposed by Pelckmans et al. (2005), Lindsten et al. (2011) and Hocking et al. (2011). The perfect recovery properties of the convex clustering model with uniformly weighted all-pairwise-differences regularization have been proved by Zhu et al. (2014) and Panahi et al. (2017). However, no theoretical guarantee has been established for the general weighted convex clustering model, where better empirical results have been observed. In the numerical optimization aspect, although algorithms like the alternating direction method of multipliers (ADMM) and the alternating minimization algorithm (AMA) have been proposed to solve the convex clustering model (Chi and Lange, 2015), it still remains very challenging to solve large-scale problems. In this paper, we establish su
AB - Clustering is a fundamental problem in unsupervised learning. Popular methods like Kmeans, may suffer from poor performance as they are prone to get stuck in its local minima. Recently, the sum-of-norms (SON) model (also known as the convex clustering model) has been proposed by Pelckmans et al. (2005), Lindsten et al. (2011) and Hocking et al. (2011). The perfect recovery properties of the convex clustering model with uniformly weighted all-pairwise-differences regularization have been proved by Zhu et al. (2014) and Panahi et al. (2017). However, no theoretical guarantee has been established for the general weighted convex clustering model, where better empirical results have been observed. In the numerical optimization aspect, although algorithms like the alternating direction method of multipliers (ADMM) and the alternating minimization algorithm (AMA) have been proposed to solve the convex clustering model (Chi and Lange, 2015), it still remains very challenging to solve large-scale problems. In this paper, we establish su
KW - Augmented Lagrangian method
KW - Conjugate gradient method
KW - Convex clustering
KW - Semismooth Newton method
KW - Unsupervised learning
UR - http://www.scopus.com/inward/record.url?scp=85105850315&partnerID=8YFLogxK
M3 - Journal article
AN - SCOPUS:85105850315
SN - 1532-4435
VL - 22
SP - 1
EP - 32
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
ER -