Reinforcement-learning-based output-feedback control of nonstrict nonlinear discrete-time systems with application to engine emission control
- PMID: 19336317
- DOI: 10.1109/TSMCB.2009.2013272
Reinforcement-learning-based output-feedback control of nonstrict nonlinear discrete-time systems with application to engine emission control
Abstract
A novel reinforcement-learning-based output adaptive neural network (NN) controller, which is also referred to as the adaptive-critic NN controller, is developed to deliver the desired tracking performance for a class of nonlinear discrete-time systems expressed in nonstrict feedback form in the presence of bounded and unknown disturbances. The adaptive-critic NN controller consists of an observer, a critic, and two action NNs. The observer estimates the states and output, and the two action NNs provide virtual and actual control inputs to the nonlinear discrete-time system. The critic approximates a certain strategic utility function, and the action NNs minimize the strategic utility function and control inputs. All NN weights adapt online toward minimization of a performance index, utilizing the gradient-descent-based rule, in contrast with iteration-based adaptive-critic schemes. Lyapunov functions are used to show the stability of the closed-loop tracking error, weights, and observer estimates. Separation and certainty equivalence principles, persistency of excitation condition, and linearity in the unknown parameter assumption are not needed. Experimental results on a spark ignition (SI) engine operating lean at an equivalence ratio of 0.75 show a significant (25%) reduction in cyclic dispersion in heat release with control, while the average fuel input changes by less than 1% compared with the uncontrolled case. Consequently, oxides of nitrogen (NO(x)) drop by 30%, and unburned hydrocarbons drop by 16% with control. Overall, NO(x)'s are reduced by over 80% compared with stoichiometric levels.
Similar articles
-
Reinforcement-learning-based dual-control methodology for complex nonlinear discrete-time systems with application to spark engine EGR operation.IEEE Trans Neural Netw. 2008 Aug;19(8):1369-88. doi: 10.1109/TNN.2008.2000452. IEEE Trans Neural Netw. 2008. PMID: 18701368
-
Control of nonaffine nonlinear discrete-time systems using reinforcement-learning-based linearly parameterized neural networks.IEEE Trans Syst Man Cybern B Cybern. 2008 Aug;38(4):994-1001. doi: 10.1109/TSMCB.2008.926607. IEEE Trans Syst Man Cybern B Cybern. 2008. PMID: 18632390
-
Neural-network-based state feedback control of a nonlinear discrete-time system in nonstrict feedback form.IEEE Trans Neural Netw. 2008 Dec;19(12):2073-87. doi: 10.1109/TNN.2008.2003295. IEEE Trans Neural Netw. 2008. PMID: 19054732
-
Issues on stability of ADP feedback controllers for dynamical systems.IEEE Trans Syst Man Cybern B Cybern. 2008 Aug;38(4):913-7. doi: 10.1109/TSMCB.2008.926599. IEEE Trans Syst Man Cybern B Cybern. 2008. PMID: 18632377 Review.
-
Higher level application of ADP: a next phase for the control field?IEEE Trans Syst Man Cybern B Cybern. 2008 Aug;38(4):901-12. doi: 10.1109/TSMCB.2008.918073. IEEE Trans Syst Man Cybern B Cybern. 2008. PMID: 18632376 Review.
Cited by
-
Safe deep reinforcement learning in diesel engine emission control.Proc Inst Mech Eng Part I J Syst Control Eng. 2023 Sep;237(8):1440-1453. doi: 10.1177/09596518231153445. Epub 2023 Feb 17. Proc Inst Mech Eng Part I J Syst Control Eng. 2023. PMID: 37692899 Free PMC article.
-
Adaptive Finite-Time-Based Neural Optimal Control of Time-Delayed Wheeled Mobile Robotics Systems.Sensors (Basel). 2024 Aug 23;24(17):5462. doi: 10.3390/s24175462. Sensors (Basel). 2024. PMID: 39275373 Free PMC article.
Publication types
MeSH terms
Substances
LinkOut - more resources
Full Text Sources
Other Literature Sources