Optimal Event-Triggered Control Using Adaptive Dynamic Programming

Optimal Event-Triggered Control Using Adaptive Dynamic Programming PDF

Author: Sarangapani Jagannathan

Publisher: CRC Press

Published: 2024-06-21

Total Pages: 348

ISBN-13: 1040049168

DOWNLOAD EBOOK →

Optimal Event-triggered Control using Adaptive Dynamic Programming discusses event triggered controller design which includes optimal control and event sampling design for linear and nonlinear dynamic systems including networked control systems (NCS) when the system dynamics are both known and uncertain. The NCS are a first step to realize cyber-physical systems (CPS) or industry 4.0 vision. The authors apply several powerful modern control techniques to the design of event-triggered controllers and derive event-trigger condition and demonstrate closed-loop stability. Detailed derivations, rigorous stability proofs, computer simulation examples, and downloadable MATLAB® codes are included for each case. The book begins by providing background on linear and nonlinear systems, NCS, networked imperfections, distributed systems, adaptive dynamic programming and optimal control, stability theory, and optimal adaptive event-triggered controller design in continuous-time and discrete-time for linear, nonlinear and distributed systems. It lays the foundation for reinforcement learning-based optimal adaptive controller use for infinite horizons. The text then: Introduces event triggered control of linear and nonlinear systems, describing the design of adaptive controllers for them Presents neural network-based optimal adaptive control and game theoretic formulation of linear and nonlinear systems enclosed by a communication network Addresses the stochastic optimal control of linear and nonlinear NCS by using neuro dynamic programming Explores optimal adaptive design for nonlinear two-player zero-sum games under communication constraints to solve optimal policy and event trigger condition Treats an event-sampled distributed linear and nonlinear systems to minimize transmission of state and control signals within the feedback loop via the communication network Covers several examples along the way and provides applications of event triggered control of robot manipulators, UAV and distributed joint optimal network scheduling and control design for wireless NCS/CPS in order to realize industry 4.0 vision An ideal textbook for senior undergraduate students, graduate students, university researchers, and practicing engineers, Optimal Event Triggered Control Design using Adaptive Dynamic Programming instills a solid understanding of neural network-based optimal controllers under event-sampling and how to build them so as to attain CPS or Industry 4.0 vision.

Adaptive Dynamic Programming: Single and Multiple Controllers

Adaptive Dynamic Programming: Single and Multiple Controllers PDF

Author: Ruizhuo Song

Publisher: Springer

Published: 2018-12-28

Total Pages: 271

ISBN-13: 9811317127

DOWNLOAD EBOOK →

This book presents a class of novel optimal control methods and games schemes based on adaptive dynamic programming techniques. For systems with one control input, the ADP-based optimal control is designed for different objectives, while for systems with multi-players, the optimal control inputs are proposed based on games. In order to verify the effectiveness of the proposed methods, the book analyzes the properties of the adaptive dynamic programming methods, including convergence of the iterative value functions and the stability of the system under the iterative control laws. Further, to substantiate the mathematical analysis, it presents various application examples, which provide reference to real-world practices.

Robust Adaptive Dynamic Programming

Robust Adaptive Dynamic Programming PDF

Author: Yu Jiang

Publisher: John Wiley & Sons

Published: 2017-05-08

Total Pages: 216

ISBN-13: 1119132649

DOWNLOAD EBOOK →

A comprehensive look at state-of-the-art ADP theory and real-world applications This book fills a gap in the literature by providing a theoretical framework for integrating techniques from adaptive dynamic programming (ADP) and modern nonlinear control to address data-driven optimal control design challenges arising from both parametric and dynamic uncertainties. Traditional model-based approaches leave much to be desired when addressing the challenges posed by the ever-increasing complexity of real-world engineering systems. An alternative which has received much interest in recent years are biologically-inspired approaches, primarily RADP. Despite their growing popularity worldwide, until now books on ADP have focused nearly exclusively on analysis and design, with scant consideration given to how it can be applied to address robustness issues, a new challenge arising from dynamic uncertainties encountered in common engineering problems. Robust Adaptive Dynamic Programming zeros in on the practical concerns of engineers. The authors develop RADP theory from linear systems to partially-linear, large-scale, and completely nonlinear systems. They provide in-depth coverage of state-of-the-art applications in power systems, supplemented with numerous real-world examples implemented in MATLAB. They also explore fascinating reverse engineering topics, such how ADP theory can be applied to the study of the human brain and cognition. In addition, the book: Covers the latest developments in RADP theory and applications for solving a range of systems’ complexity problems Explores multiple real-world implementations in power systems with illustrative examples backed up by reusable MATLAB code and Simulink block sets Provides an overview of nonlinear control, machine learning, and dynamic control Features discussions of novel applications for RADP theory, including an entire chapter on how it can be used as a computational mechanism of human movement control Robust Adaptive Dynamic Programming is both a valuable working resource and an intriguing exploration of contemporary ADP theory and applications for practicing engineers and advanced students in systems theory, control engineering, computer science, and applied mathematics.

Adaptive Dynamic Programming with Applications in Optimal Control

Adaptive Dynamic Programming with Applications in Optimal Control PDF

Author: Derong Liu

Publisher: Springer

Published: 2018-07-13

Total Pages: 0

ISBN-13: 9783319844978

DOWNLOAD EBOOK →

This book covers the most recent developments in adaptive dynamic programming (ADP). The text begins with a thorough background review of ADP making sure that readers are sufficiently familiar with the fundamentals. In the core of the book, the authors address first discrete- and then continuous-time systems. Coverage of discrete-time systems starts with a more general form of value iteration to demonstrate its convergence, optimality, and stability with complete and thorough theoretical analysis. A more realistic form of value iteration is studied where value function approximations are assumed to have finite errors. Adaptive Dynamic Programming also details another avenue of the ADP approach: policy iteration. Both basic and generalized forms of policy-iteration-based ADP are studied with complete and thorough theoretical analysis in terms of convergence, optimality, stability, and error bounds. Among continuous-time systems, the control of affine and nonaffine nonlinear systems is studied using the ADP approach which is then extended to other branches of control theory including decentralized control, robust and guaranteed cost control, and game theory. In the last part of the book the real-world significance of ADP theory is presented, focusing on three application examples developed from the authors’ work: • renewable energy scheduling for smart power grids;• coal gasification processes; and• water–gas shift reactions. Researchers studying intelligent control methods and practitioners looking to apply them in the chemical-process and power-supply industries will find much to interest them in this thorough treatment of an advanced approach to control.

Adaptive Dynamic Programming for Control

Adaptive Dynamic Programming for Control PDF

Author: Huaguang Zhang

Publisher: Springer Science & Business Media

Published: 2012-12-14

Total Pages: 432

ISBN-13: 144714757X

DOWNLOAD EBOOK →

There are many methods of stable controller design for nonlinear systems. In seeking to go beyond the minimum requirement of stability, Adaptive Dynamic Programming in Discrete Time approaches the challenging topic of optimal control for nonlinear systems using the tools of adaptive dynamic programming (ADP). The range of systems treated is extensive; affine, switched, singularly perturbed and time-delay nonlinear systems are discussed as are the uses of neural networks and techniques of value and policy iteration. The text features three main aspects of ADP in which the methods proposed for stabilization and for tracking and games benefit from the incorporation of optimal control methods: • infinite-horizon control for which the difficulty of solving partial differential Hamilton–Jacobi–Bellman equations directly is overcome, and proof provided that the iterative value function updating sequence converges to the infimum of all the value functions obtained by admissible control law sequences; • finite-horizon control, implemented in discrete-time nonlinear systems showing the reader how to obtain suboptimal control solutions within a fixed number of control steps and with results more easily applied in real systems than those usually gained from infinite-horizon control; • nonlinear games for which a pair of mixed optimal policies are derived for solving games both when the saddle point does not exist, and, when it does, avoiding the existence conditions of the saddle point. Non-zero-sum games are studied in the context of a single network scheme in which policies are obtained guaranteeing system stability and minimizing the individual performance function yielding a Nash equilibrium. In order to make the coverage suitable for the student as well as for the expert reader, Adaptive Dynamic Programming in Discrete Time: • establishes the fundamental theory involved clearly with each chapter devoted to a clearly identifiable control paradigm; • demonstrates convergence proofs of the ADP algorithms to deepen understanding of the derivation of stability and convergence with the iterative computational methods used; and • shows how ADP methods can be put to use both in simulation and in real applications. This text will be of considerable interest to researchers interested in optimal control and its applications in operations research, applied mathematics computational intelligence and engineering. Graduate students working in control and operations research will also find the ideas presented here to be a source of powerful methods for furthering their study.

Adaptive Dynamic Programming

Adaptive Dynamic Programming PDF

Author: Jiayue Sun

Publisher: Springer Nature

Published: 2023-10-14

Total Pages: 144

ISBN-13: 9819959292

DOWNLOAD EBOOK →

This open access book focuses on the practical application of Adaptive Dynamic Programming (ADP) in chemotherapy drug delivery, taking into account clinical variables and real-time data. ADP's ability to adapt to changing conditions and make optimal decisions in complex and uncertain situations makes it a valuable tool in addressing pressing challenges in healthcare and other fields. As optimization technology evolves, we can expect to see even more sophisticated and powerful solutions emerge.

Adaptive Dynamic Programming with Applications in Optimal Control

Adaptive Dynamic Programming with Applications in Optimal Control PDF

Author: Derong Liu

Publisher: Springer

Published: 2017-01-04

Total Pages: 594

ISBN-13: 3319508156

DOWNLOAD EBOOK →

This book covers the most recent developments in adaptive dynamic programming (ADP). The text begins with a thorough background review of ADP making sure that readers are sufficiently familiar with the fundamentals. In the core of the book, the authors address first discrete- and then continuous-time systems. Coverage of discrete-time systems starts with a more general form of value iteration to demonstrate its convergence, optimality, and stability with complete and thorough theoretical analysis. A more realistic form of value iteration is studied where value function approximations are assumed to have finite errors. Adaptive Dynamic Programming also details another avenue of the ADP approach: policy iteration. Both basic and generalized forms of policy-iteration-based ADP are studied with complete and thorough theoretical analysis in terms of convergence, optimality, stability, and error bounds. Among continuous-time systems, the control of affine and nonaffine nonlinear systems is studied using the ADP approach which is then extended to other branches of control theory including decentralized control, robust and guaranteed cost control, and game theory. In the last part of the book the real-world significance of ADP theory is presented, focusing on three application examples developed from the authors’ work: • renewable energy scheduling for smart power grids;• coal gasification processes; and• water–gas shift reactions. Researchers studying intelligent control methods and practitioners looking to apply them in the chemical-process and power-supply industries will find much to interest them in this thorough treatment of an advanced approach to control.

Reinforcement Learning and Dynamic Programming Using Function Approximators

Reinforcement Learning and Dynamic Programming Using Function Approximators PDF

Author: Lucian Busoniu

Publisher: CRC Press

Published: 2017-07-28

Total Pages: 280

ISBN-13: 1439821097

DOWNLOAD EBOOK →

From household appliances to applications in robotics, engineered systems involving complex dynamics can only be as effective as the algorithms that control them. While Dynamic Programming (DP) has provided researchers with a way to optimally solve decision and control problems involving complex dynamic systems, its practical value was limited by algorithms that lacked the capacity to scale up to realistic problems. However, in recent years, dramatic developments in Reinforcement Learning (RL), the model-free counterpart of DP, changed our understanding of what is possible. Those developments led to the creation of reliable methods that can be applied even when a mathematical model of the system is unavailable, allowing researchers to solve challenging control problems in engineering, as well as in a variety of other disciplines, including economics, medicine, and artificial intelligence. Reinforcement Learning and Dynamic Programming Using Function Approximators provides a comprehensive and unparalleled exploration of the field of RL and DP. With a focus on continuous-variable problems, this seminal text details essential developments that have substantially altered the field over the past decade. In its pages, pioneering experts provide a concise introduction to classical RL and DP, followed by an extensive presentation of the state-of-the-art and novel methods in RL and DP with approximation. Combining algorithm development with theoretical guarantees, they elaborate on their work with illustrative examples and insightful comparisons. Three individual chapters are dedicated to representative algorithms from each of the major classes of techniques: value iteration, policy iteration, and policy search. The features and performance of these algorithms are highlighted in extensive experimental studies on a range of control applications. The recent development of applications involving complex systems has led to a surge of interest in RL and DP methods and the subsequent need for a quality resource on the subject. For graduate students and others new to the field, this book offers a thorough introduction to both the basics and emerging methods. And for those researchers and practitioners working in the fields of optimal and adaptive control, machine learning, artificial intelligence, and operations research, this resource offers a combination of practical algorithms, theoretical analysis, and comprehensive examples that they will be able to adapt and apply to their own work. Access the authors' website at www.dcsc.tudelft.nl/rlbook/ for additional material, including computer code used in the studies and information concerning new developments.

Proceedings of 2021 5th Chinese Conference on Swarm Intelligence and Cooperative Control

Proceedings of 2021 5th Chinese Conference on Swarm Intelligence and Cooperative Control PDF

Author: Zhang Ren

Publisher: Springer Nature

Published: 2022-07-29

Total Pages: 1902

ISBN-13: 9811939985

DOWNLOAD EBOOK →

This book includes original, peer-reviewed research papers from the 2021 5th Chinese Conference on Swarm Intelligence and Cooperative Control (CCSICC2021), held in Shenzhen, China on January 19-22, 2022. The topics covered include but are not limited to: reviews and discussions of swarm intelligence, basic theories on swarm intelligence, swarm communication and networking, swarm perception, awareness and location, swarm decision and planning, cooperative control, cooperative guidance, swarm simulation and assessment. The papers showcased here share the latest findings on theories, algorithms and applications in swarm intelligence and cooperative control, making the book a valuable asset for researchers, engineers, and university students alike.

Advanced Optimal Control and Applications Involving Critic Intelligence

Advanced Optimal Control and Applications Involving Critic Intelligence PDF

Author: Ding Wang

Publisher: Springer Nature

Published: 2023-01-21

Total Pages: 283

ISBN-13: 9811972915

DOWNLOAD EBOOK →

This book intends to report new optimal control results with critic intelligence for complex discrete-time systems, which covers the novel control theory, advanced control methods, and typical applications for wastewater treatment systems. Therein, combining with artificial intelligence techniques, such as neural networks and reinforcement learning, the novel intelligent critic control theory as well as a series of advanced optimal regulation and trajectory tracking strategies are established for discrete-time nonlinear systems, followed by application verifications to complex wastewater treatment processes. Consequently, developing such kind of critic intelligence approaches is of great significance for nonlinear optimization and wastewater recycling. The book is likely to be of interest to researchers and practitioners as well as graduate students in automation, computer science, and process industry who wish to learn core principles, methods, algorithms, and applications in the field of intelligent optimal control. It is beneficial to promote the development of intelligent optimal control approaches and the construction of high-level intelligent systems.