(227h) Global Optimization for Clustering Problems with 200,000 Samples
AIChE Annual Meeting
2021
2021 Annual Meeting
Computing and Systems Technology Division
Advances in global optimization
Tuesday, November 9, 2021 - 10:13am to 10:32am
The MSSC task is rarely investigated in the literature through deterministic optimization to find its global optimal value. One reason is that a classic global optimization algorithm based on branch and bound (BB) scheme needs to perform branching on all binary variables, and the number of binary variables increases linearly as the number of data samples. Therefore, off-the-shelf global solvers cannot solve clustering problems even with hundreds of data points. In this presentation, we proposed a tailed reduced-space BB algorithm and designed four approaches to construct lower and upper bounds at each node in the BB scheme. One key advantage of this reduced-space algorithm is that it only needs to perform branching on the centers of clusters to guarantee convergence, and the size of centers is independent of the number of data samples. Another critical property of this algorithm is that both lower bounds and upper bounds can be obtained without solving any optimization problems. These two properties enable our algorithm to be scalable to the dataset with hundreds of thousands of data points. We performed numerical experiments on both synthetic and real-world datasets and compared our proposed algorithms with the off-the-shelf global optimal solvers and classical local optimal algorithms. The results reveal a strong performance and scalability of our algorithm. For example, the parallel version of our algorithm can solve the MSSC problem to an optimality gap of 1% on a synthetic dataset of 210,000 samples (i.e., 100 times larger than the state-of-the-art method [4] in the literature), using 100 CPU cores.
[1] Jain, A. K. Data clustering: 50 years beyond k-means. Pattern recognition letters, 31(8):651â666, 2010.
[2] Lloyd, S. Least squares quantization in pcm. IEEE transactions on information theory, 28(2):129â137, 1982.
[3] Xu, J. and Lange, K. Power k-means clustering. In International Conference on Machine Learning, pp. 6921â6931, 2019.
[4]. Aloise, D., Hansen, P. and Liberti, L. An improved column generation algorithm for minimum sum-of-squares clustering. Mathematical Programming, 131(1):195-220, 2012.