(402b) Updated LEAPS2 for Surrogate Recommendation
AIChE Annual Meeting
2020
2020 Virtual AIChE Annual Meeting
Computing and Systems Technology Division
Advances in Machine Learning and Intelligent Systems I
Tuesday, November 17, 2020 - 8:15am to 8:30am
In this work, we modified and broadened the scope of LEAPS2 in several significant ways. First, we incorporated noisy and real-world data sets to address a key challenge in surrogate modeling. Second, we added one more metric for surrogate selection, namely a complexity-based metric called AIC weight. This metric provides an alternative for surrogate selection when splitting the dataset into train/test sets is not feasible. Third, we essentially revamped the attribute set of LEAPS2 to use only those attributes that quantify the underlying features of âdata-distributionâ, rather than the data itself. Some of these new attributes quantify the degree and variations of non-linearity in the data, asymmetry, and flatness of response with respect to standard distribution. Thus, we now have fewer (11 vs 14) but intuitively more appealing attributes in LEAPS2. Fourth, we have improved the surrogate recommendation strategy by developing simple heuristics. Finally, we have updated our surrogate pool by adding 10 new surrogates in LEAPS2. Our improved LEAPS2 framework was evaluated with respect to the two metrics (Garud et al. 2018), namely âTotal Degree of Successâ (TDoS) that quantifies the success in recommending the best surrogates, and âTotal Coefficient of Rewardâ (TCoR) that combines the success and computational savings in a single score. The new framework gives a TDoS = 91% and a TCoR = 42% for the error-based metric, and TDoS = 83% but a much higher TCoR = 63% for AIC weight on test data. However, they improved during the learning process. We tested the new framework on two case studies with real data, one on a compressor, and the other on COVID-19 data. In both cases, our improved LEAPS2 achieved a TDoS of 100%. This framework acts as a smart tool for surrogate selection to model complex physical systems.
References:
Cozad, Alison, and Nikolaos V. Sahinidis. 2018. âA Global MINLP Approach to Symbolic Regression.â Mathematical Programming 170 (1): 97â119. https://doi.org/10.1007/s10107-018-1289-x.
Cozad, Alison, Nikolaos V. Sahinidis, and David C. Miller. 2014. âLearning Surrogate Models for Simulation-Based Optimization.â AIChE Journal 60 (6): 2211â27. https://doi.org/10.1002/aic.14418.
Cui, Can, Mengqi Hu, Jeffery D. Weir, and Teresa Wu. 2016. âA Recommendation System for Meta-Modeling: A Meta-Learning Based Approach.â Expert Systems with Applications 46 (March): 33â44. https://doi.org/10.1016/j.eswa.2015.10.021.
Davis, Sarah E., Selen Cremaschi, and Mario R. Eden. 2018. âEfficient Surrogate Model Development: Impact of Sample Size and Underlying Model Dimensions.â In Computer Aided Chemical Engineering, 44:979â84. Elsevier. https://doi.org/10.1016/B978-0-444-64241-7.50158-0.
Garud, Sushant S., Iftekhar A. Karimi, and Markus Kraft. 2018. âLEAPS2: Learning Based Evolutionary Assistive Paradigm for Surrogate Selection.â Computers & Chemical Engineering 119 (November): 352â70. https://doi.org/10.1016/j.compchemeng.2018.09.008.
Koza, JohnR. 1994. âGenetic Programming as a Means for Programming Computers by Natural Selection.â Statistics and Computing 4 (2). https://doi.org/10.1007/BF00175355.
Lessmann, Stefan, Robert Stahlbock, and Sven F Crone. 2006. âGenetic Algorithms for Support Vector Machine Model Selection,â 7.
Rad, Hossein Izadi, Ji Feng, and Hitoshi Iba. 2018. âGP-RVM: Genetic Programing-Based Symbolic Regression Using Relevance Vector Machine.â ArXiv:1806.02502 [Cs], August. http://arxiv.org/abs/1806.02502.
Streeter, Matthew, and Lee A Becker. 2003. âAutomated Discovery of Numerical Approximation Formulae via Genetic Programming,â 32.