(457e) MPC Is a Decision Tree, and We Know How to Optimize It | AIChE

(457e) MPC Is a Decision Tree, and We Know How to Optimize It

Authors 

Cao, Y., The University of British Columbia
Model Predictive Control (MPC) has gained extensive adoption in the process industry due to its superior control performance and its ability to handle constraints [1]. However, the real-time computational demands of solving the optimization problem often restrict MPC implementation in systems characterized by fast dynamics or limited computational resources.

One strategy to address this challenge is to compute or approximate optimal control law offline. Explicit MPC, for instance, computes piecewise affine functions of the system states as control laws for MPC problems with linear plant models [2], [3]. Nonetheless, the scalability of Explicit MPC to large-scale systems faces obstacles due to the exponential growth of partitions with increasing system states and control horizons. An alternative approach involves approximating the MPC control law via neural networks [4], [5], [6]. However, one limitation of the neural network control law is its inherent lack of interpretability, making it difficult for engineers to comprehend the decision-making process, which is crucial in safety-critical systems.

In this paper, we propose to train oblique decision trees (DTs) with linear predictions to approximate the MPC control laws offline, followed by utilizing the DT model as the online controller. This method is based on the observation that MPC control law can be viewed as a DT. As illustrated in Figure 1, DTs with oblique splits at the branch nodes and linear predictions at the leaf nodes exhibit the same formulations as piecewise affine functions, akin to explicit MPC control law. One notable advantage of the DT control law lies in its interpretability; it follows if-else rules, akin to human reasoning, and is easily comprehensible by human decision-makers. Furthermore, scalability is another key benefit, as the control law can be derived through a data-driven approach, with the training dataset generated by simulating an ideal MPC in parallel.

One key challenge of the proposed method is computing the most accurate DT model from a dataset, which is a mixed-integer optimization problem [7]. To tackle this challenge, we introduce a novel gradient-based algorithm tailored for oblique decision tree training. This approach involves approximating the gradients of indicator functions using scaled sigmoid functions, enabling effective utilization of existing machine learning packages on GPUs.

We validate the effectiveness of the proposed method through various case studies, demonstrating its capability to accurately approximate the MPC control law while significantly reducing the online computational time. This approach matches the performance of the ground truth control law for linear MPC and accurately approximates the control law for nonlinear MPC.

[1] S. J. Qin and T. A. Badgwell, “A survey of industrial model predictive control technology,” Control Engineering Practice, vol. 11, no. 7, pp. 733–764, Jul. 2003, doi: 10.1016/S0967-0661(02)00186-7.

[2] A. Bemporad, M. Morari, V. Dua, and E. N. Pistikopoulos, “The explicit linear quadratic regulator for constrained systems,” Automatica, vol. 38, no. 1, pp. 3–20, 2002, doi: 10.1016/S0005-1098(01)00174-1.

[3] A. Alessio and A. Bemporad, “A survey on explicit model predictive control,” Nonlinear Model Predictive Control: Towards New Challenging Applications, pp. 345–369, 2009, doi: 10.1007/978-3-642-01094-1_29.

[4] T. Parisini and R. Zoppoli, “A receding-horizon regulator for nonlinear systems and a neural approximation,” Automatica, vol. 31, no. 10, pp. 1443–1451, Oct. 1995, doi: 10.1016/0005-1098(95)00044-W.

[5] P. Kumar, J. B. Rawlings, and S. J. Wright, “Industrial, large-scale model predictive control with structured neural networks,” Computers & Chemical Engineering, vol. 150, p. 107291, Jul. 2021, doi: 10.1016/j.compchemeng.2021.107291.

[6] Y. Li, K. Hua, and Y. Cao, “Using stochastic programming to train neural network approximation of nonlinear MPC laws,” Automatica, vol. 146, p. 110665, Dec. 2022, doi: 10.1016/j.automatica.2022.110665..

[7] K. Hua, J. Ren, and Y. Cao, “A Scalable Deterministic Global Optimization Algorithm for Training Optimal Decision Tree,” presented at the Advances in Neural Information Processing Systems, Oct. 2022.