(295h) Data-Driven State Observation Via Online Optimization of Chen-Fliess Series
AIChE Annual Meeting
2023
2023 AIChE Annual Meeting
Computing and Systems Technology Division
Advances in Process Control II
Wednesday, November 8, 2023 - 10:06am to 10:24am
However, the direct solution of a KKL observer requires the solution of a PDE system to find a transformation of states such that the immersed dynamics is linear and driven by the outputs. This is analytically infeasible in general and numerically difficult under high dimensions. Hence, the attention of recent research has largely been on approximating the (inverse) immersion mapping as neural networks and training them with process data (e.g., [3, 4]). Yet, the neural network training can require massive data and the problem itself is nonconvex, subject to the effect of initialization and local optima. The neural network solutions also do not provide a guaranteed performance.
This work proposes a computationally efficient and performance-provable approach for the data-driven state observation of nonlinear systems. Specifically,
- The observer, as a map from the plant outputs to the state estimates, can be expanded on every small time interval as a Chen-Fliess series. Each term of the series contains a recursive integral of the output signal, multiplied by a coefficient, which is the corresponding Lie derivative evaluated at the initial time [5, Section 3.1]. Hence, on this interval, the Chen-Fliess seriesâ coefficients can be obtained through linear regression.
- Due to the local convergence property of the Chen-Fliess series, as time proceeds, the regression should be recursive. In continuous time, this work proposes to impose a gradient flow on the least-squares estimate, thus forming an online optimization algorithm [6] for Chen-Fliess series regression. If the true states are hypothetically measurable, the resulting dynamic regret, namely the integrated squares of observation error, is proved to be finite L2-gain stable with respect to the total variation of true states.
- In realistic cases where the true state trajectories are not available, the immersed linear dynamics of the KKL observer is first assigned so that the transformed states can be computed. The online optimization of Chen-Fliess series is then conducted on the transformed state space, and dimensionality reduction is adopted to recover the estimates of the equivalents of the states (in the sense of a diffeomorphism).
The proposed approach is demonstrated by applications to a benchmark nonlinear process. It is worth noting that the proposed approach intrinsically does not use the model information (i.e., governing equations) and is thus model-free. Therefore, it is promising for combination with model-free control strategies (e.g., reinforcement learning and dissipativity learning control) [7].
References
[1] Kazantzis, N., & Kravaris, C. (1998). Systems & Control Letters, 34, 241-247.
[2] Bernard, P., Andrieu, V., & Astolfi, D. (2022). Annual Reviews in Control, 53, 224-248.
[3] Ramos, L. D. C., et al. (2020, December). IEEE Conference on Decision and Control (pp. 5435-5442).
[4] Niazi, M. U. B., et al. (2022). arXiv:2210.01476.
[5] Isidori, A. (1995). Nonlinear control systems. Springer.
[6] Hazan, E. (2016). Foundations and Trends in Optimization, 2, 157-325.
[7] Tang, W., & Daoutidis, P. (2022, June). American Control Conference (pp. 1048-1064).