(674i) Accelerated Modeling of Various Chemical Processes Using Meta-Learning-Based Foundation Models: A Few-Shot Learning Approach Using Reptile
AIChE Annual Meeting
2024
2024 AIChE Annual Meeting
Computing and Systems Technology Division
10B: AI/ML Modeling, Optimization and Control Applications I
Thursday, October 31, 2024 - 2:38pm to 2:54pm
To address the limitations of transfer learning, we will develop ONE universal neural network that is capable of swiftly adapting to any new task, such as modeling system dynamics, following the idea of foundation models. We employ a meta-learning technique, Reptile [4] that is proposed by OpenAI to optimize neural network weights, to facilitate easy adaptation to new tasks with just a few shots from those tasks. While meta-learning techniques are commonly used in few-shot image classification problems, the application to time-series regression is still in its infancy [5]. In this study, we consider an example of neural network modeling of nonlinear chemical reactors, and develop a universal neural network to encompass various chemical reactors such as continuous stirred tank reactors (CSTRs), batch reactors (BRs), and plug flow reactors (PFRs). To validate the efficacy of the Neural Network (NN)-based Reptile in few-shot learning for various chemical reactors, two sets of simulations on nonlinear processes were conducted. Firstly, we trained an RNN-based Reptile with 1,000 CSTRs of different parameters (e.g., volume, reaction rates, inlet flow rates, etc.) and demonstrated its ability to enable few-shot learning on unseen CSTRs, achieving similar accuracy with as few as 10 shots compared to training with a large number of samples. Secondly, we trained an NN-based Reptile with 1,000 CSTRs, BRs, and PFRs of different parameters, showcasing their effectiveness in few-shot learning on unseen chemical reactions. Again, with as few as 10 shots, it achieved similar accuracy as training with sufficiently large number of samples. Notably, the transfer learning-based approach failed to achieve satisfactory performance in few-shot learning under both settings.
References:
[1] Z. Wu, A. Tran, D. Rincon, and P. D. Christofides, âMachine learningâbased predictive control of nonlinear processes. Part I: theory,â AIChE J., vol. 65, no. 11, p. e16729, 2019.
[2] Z. Wu, A. Tran, D. Rincon, and P. D. Christofides, âMachineâlearningâbased predictive control of nonlinear processes. Part II: Computational implementation,â AIChE J., vol. 65, no. 11, p. e16734, 2019.
[3] F. Zhuang et al., âA comprehensive survey on transfer learning,â Proc. IEEE, vol. 109, no. 1, pp. 43â76, 2020.
[4] A. Nichol and J. Schulman, âReptile: a scalable metalearning algorithm,â ArXiv Prepr. ArXiv180302999, vol. 2, no. 3, p. 4, 2018.
[5] T. Hospedales, A. Antoniou, P. Micaelli, and A. Storkey, âMeta-learning in neural networks: A survey,â IEEE Trans. Pattern Anal. Mach. Intell., vol. 44, no. 9, pp. 5149â5169, 2021.