2024 AIChE Annual Meeting
(676g) Development of Steady-State and Dynamic Mass-Energy Constrained Neural Network Models Using Noisy Temporal Data for Dynamic Optimization of Distributed Chemical Systems
Authors
Recent years have seen the development of a relatively new branch of supervised data-driven modeling of nonlinear systems known as physics-informed neural networks3 (PINNs) where typically desired physics constraints are imposed by penalizing the objective function in typical NN training algorithms, thus serving as a soft constraint only during the inverse problem. But in most chemical engineering applications, it is desired that certain mathematical relationships are exactly satisfied, which cannot be guaranteed by these PINNs4. Moreover, most hybrid mechanistic approaches focused on exactly conserving mass and/or energy of a system require rigorous understanding of the process for formulating the physics-based differential or algebraic constraints and hence become system-specific5. In this work, a novel class of network models in proposed, namely the Mass-Energy Constrained Neural Networks (MECNNs), that exactly satisfies the mass and energy constraints, expressed in terms of the species molar/atom or enthalpy balance equations and posed as equality constraints in the nonlinear parameter estimation problem. The developed algorithms for both inverse and forward problems provide sufficient flexibility to model any distributed dynamic nonlinear chemical process system involving addition / removal of mass and/or heat as well as transformation of energy from one form to another. The proposed approaches can also accommodate nonlinear transformation for constraint satisfaction, even when there is no close-form equation available for the same (e.g., constraint satisfaction may require solution of an optimization problem). Efficient algorithms are developed for optimal network synthesis and parameter estimation involving the incorporation of local measurements especially while modeling distributed process systems. The proposed algorithms for solving both the inverse and forward problems are tested by using noisy steady-state and dynamic training data. Furthermore, most data-driven approaches for modeling complex nonlinear dynamic systems with respect to available measurements may not provide any information about the ‘true’ data. The optimal MECNNs developed in this work yields the system truth with minimum bias when the noise in the training data is represented by a Gaussian distribution with time-invariant as well as time-varying mean.
Unlike steady-state, developing a fully data-driven dynamic modeling approach by exactly satisfying the mass and energy balance equations can be significantly challenging, since conservation of mass/energy during transience is difficult to satisfy, in general, due to insufficient information about the holdup of a system. Efficient algorithms are also developed in this work for estimating optimal parameters for dynamic MECNNs represented by hybrid series/parallel all-nonlinear static-dynamic neural network models (developed as part of our previous work6), which have been shown to perform significantly superior to many existing state-of-the-art approaches in terms of both computational expense as well as predictive accuracy, while modeling many complex nonlinear dynamic chemical process systems6. The proposed structures and algorithms are applied to model three nonlinear dynamic chemical processes, namely a distributed superheater system, a nonisothermal continuous stirred tank reactor (CSTR) involving exothermic reaction, i.e., generation of heat within the system and an electrically heated tubular reactor where one form of energy (electric) gets transformed to another (heat). It has been observed that the outputs from the MECNN exactly satisfy mass and energy conservation, even though the data used for training the network violate the same. In addition, the model output exactly satisfy mass and energy conservation thus making them particularly attractive for dynamic optimization of chemical processes.
References
- Su, H.-T., Bhat, N., Minderman, P. A. & McAvoy, T. J. Integrating Neural Networks With First Principles Models for Dynamic Modeling. Dynamics and Control of Chemical Reactors, Distillation Columns and Batch Processes (IFAC, 1993). doi:10.1016/b978-0-08-041711-0.50054-4.
- Venkatasubramanian, V. The promise of artificial intelligence in chemical engineering: Is it here, finally? AIChE J. 65, 466–478 (2019).
- Raissi, M., Perdikaris, P. & Karniadakis, G. E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 378, 686–707 (2019).
- Carranza-Abaid, A. & Jakobsen, J. P. Neural network programming: Integrating first principles into machine learning models. Comput. Chem. Eng. 163, 107858 (2022).
- Rajulapati, L., Chinta, S., Shyamala, B. & Rengaswamy, R. Integration of machine learning and first principles models. AIChE J. 68, 1–17 (2022).
- Mukherjee, A. & Bhattacharyya, D. Hybrid Series/Parallel All-Nonlinear Dynamic-Static Neural Networks: Development, Training, and Application to Chemical Processes. Ind. Eng. Chem. Res. 62, 3221–3237 (2023).