(251g) Nonlinear Programming on GPUs for AI-Assisted Decision-Making and Beyond
AIChE Annual Meeting
2024
2024 AIChE Annual Meeting
Computing and Systems Technology Division
10C: Advances in Optimization
Tuesday, October 29, 2024 - 10:06am to 10:27am
In this presentation, we will discuss this exciting new frontier of mathematical programming on GPUs, focusing on the nonlinear optimization aspect. Despite the impressive performance of GPUs in deep learning, their usage in solving large-scale nonlinear programming has been traditionally limited due to the practical difficulties in implementing sparse automatic differentiation and sparse matrix factorization routines. In particular, the previous generations of GPU-accelerated solutions tools have been significantly slower than the alternative tools on CPUs [4]. In our recent approaches [3], we have overcome the previous challenges with two novel strategies: single instruction, multiple data (SIMD) abstractions for NLPs, which addresses the issue of sparse automatic differentiation, and a condensed-space interior point method with inequality relaxation, which resolves the challenges associated with sparse linear algebra. Our empirical results on AC optimal power flow problems highlight the considerable advantages of utilizing GPUs. Our software implementations, ExaModels and MadNLP, when executed on GPUs, deliver performance gains exceeding 20 times that of the fastest CPU-based alternatives. We also compare this approach with the previously proposed Hybrid KKT [5] and reduced-space approaches [6] and discuss the current limitation of the GPU solvers---the numerical instability due to ill-conditioning and limited portability.
Additionally, we will present our initial empirical results on solving AI-assisted decision-making problems---more specifically, neural-network-constrained optimization problems---using GPU solvers. There has been growing interest in these problems, as neural network surrogate models are increasingly employed in different applications [7]. These optimization problems embed large-scale neural networks in their objective and constraints, and the complexity of neural net surrogate models makes their solution computationally challenging. We will present numerical results highlighting the effectiveness of the GPU-accelerated solution approaches in handling the complexity of neural network surrogate models.
[1] M. Schubiger, G. Banjac, and J. Lygeros, âGPU acceleration of ADMM for large-scale quadratic programming,â Journal of Parallel and Distributed Computing, vol. 144, pp. 55â67, Oct. 2020, doi: 10.1016/j.jpdc.2020.05.021.
[2] H. Lu et al., âcuPDLP-C: A Strengthened Implementation of cuPDLP for Linear Programming by C language.â arXiv, Jan. 07, 2024. doi: 10.48550/arXiv.2312.14832.
[3] S. Shin, F. Pacaud, and M. Anitescu, âAccelerating Optimal Power Flow with GPUs: SIMD Abstraction of Nonlinear Programs and Condensed-Space Interior-Point Methods.â arXiv, Jul. 31, 2023. doi: 10.48550/arXiv.2307.16830.
[4] B. Tasseff, C. Coffrin, A. Wächter, and C. Laird, âExploring Benefits of Linear Solver Parallelism on Modern Nonlinear Optimization Applications.â arXiv, Sep. 17, 2019. doi: 10.48550/arXiv.1909.08104.
[5] S. Regev et al., âHyKKT: a hybrid direct-iterative method for solving KKT linear systems,â Optimization Methods and Software, vol. 38, no. 2, pp. 332â355, Mar. 2023, doi: 10.1080/10556788.2022.2124990.
[6] F. Pacaud, S. Shin, M. Schanen, D. A. Maldonado, and M. Anitescu, âAccelerating Condensed Interior-Point Methods on SIMD/GPU Architectures,â J Optim Theory Appl, Feb. 2023, doi: 10.1007/s10957-022-02129-5.
[7] F. Ceccon et al., âOMLT: Optimization & Machine Learning Toolkit,â Journal of Machine Learning Research, vol. 23, no. 349, pp. 1â8, 2022.