2021 Annual Meeting
(303e) Ensuring Convexity in Constrained Optimal Control Using Input Convex Neural Networks
Authors
Input convex neural networks (ICNNs) are a family of deep learning models where the outputs are constructed to be convex functions of the inputs [4]. Similar to ordinary ANNs, ICNNs can efficiently approximate all continuous Lipschitz convex functions [5]. Additionally, ICNNs can be efficiently trained on GPUs using existing deep learning packages such as PyTorch and TensorFlow. By modeling systems using ICNNs, optimal control problems on top of the system models can be solved as convex optimization problems, leading to improved performance and robustness.
Current work discusses using ICNNs for constrained optimal control of nonlinear systems. This strategy uses Model Predictive Control (MPC) on top of ICNN-based system models to dynamically optimize system performance. The strategy consists of three basic steps: Firstly, the relationships that can be approximated by convex function are identified using field knowledge or simulation. Then, process data are used to train ICNN-based system and constraint models. Lastly, the system and constraint models are integrated into the MPC framework. Because both system dynamics and constraints are modeled using ICNNs, their Jacobians can be easily calculated using backpropagation, and the MPC problems can be solved efficiently using sequential quadratic programming (SQP).
The proposed approach will be illustrated through a case study based on the Van de Vusse reactor [6], which exhibits input multiplicity and nonminimum phase behavior. The simulation results demonstrate improved economic yield compared with normal ANNs. Additionally, the input convexity formulation is compared with simple regularization techniques, and unique benefits such as improved data efficiency and robustness of the proposed formulation are shown. By explicitly incorporating prior knowledge about convexity, this framework provides a good balance between the universal approximation power of deep learning and the computational feasibility required by control and optimization.
References
[1] M. Kuure-Kinsey, R. Cutright, B.W. Bequette, Computationally efficient neural predictive control based on a feedforward architecture, Ind. Eng. Chem. Res. 45 (2006) 8575â8582.
[2] N. Bhat, T.J. McAvoy, Use of neural nets for dynamic modeling and control of chemical process systems, Comput. Chem. Eng. 14 (1990) 573â582.
[3] K. Kawaguchi, Deep Learning without Poor Local Minima, in: D.D. Lee, M. Sugiyama, U. V Luxburg, I. Guyon, R. Garnett (Eds.), Adv. Neural Inf. Process. Syst. 29, Curran Associates, Inc., 2016: pp. 586â594.
[4] B. Amos, L. Xu, J. Zico Kolter, J.Z. Kolter, Input Convex Neural Networks, in: Proc. 34th Int. Conf. Mach. Learn. - Vol. 70, JMLR.org, 2017: pp. 146â155.
[5] Y. Chen, Y. Shi, B. Zhang, Optimal Control Via Neural Networks: A Convex Approach, in: Int. Conf. Learn. Represent., 2018.
[6] B.W. Bequette, Process control: modeling, design, and simulation, Prentice Hall Professional, Upper Saddle River, NJ, United States, 2003.