The problem of H∞ state feedback control of affine nonlinear discrete-time systems with unknown dynamics is investigated in this paper. An online adaptive policy learning algorithm (APLA) based on adaptive dynamic programming (ADP) is proposed for learning in real-time the solution to the Hamilton-Jacobi-Isaacs (HJI) equation, which appears in the H∞ control problem. In the proposed algorithm, three neural networks (NNs) are utilized to find suitable approximations of the optimal value function and the saddle point feedback control and disturbance policies. Novel weight updating laws are given to tune the critic, actor, and disturbance NNs simultaneously by using data generated in real-time along the system trajectories. Considering NN approximation errors, we provide the stability analysis of the proposed algorithm with Lyapunov approach. Moreover, the need of the system input dynamics for the proposed algorithm is relaxed by using a NN identification scheme. Finally, simulation examples show the effectiveness of the proposed algorithm.
In this paper, a novel theoretic formulation based on adaptive dynamic programming (ADP) is developed to solve online the optimal tracking problem of the continuous-time linear system with unknown dynamics. First, the original system dynamics and the reference trajectory dynamics are transformed into an augmented system. Then, under the same performance index with the original system dynamics, an augmented algebraic Riccati equation is derived. Furthermore, the solutions for the optimal control problem of the augmented system are proven to be equal to the standard solutions for the optimal tracking problem of the original system dynamics. Moreover, a new online algorithm based on the ADP technique is presented to solve the optimal tracking problem of the linear system with unknown system dynamics. Finally, simulation results are given to verify the effectiveness of the theoretic results.
In this paper, we aim to solve the finite-horizon optimal control problem for a class of non-linear discrete-time switched systems using adaptive dynamic programming(ADP) algorithm. A new -optimal control scheme based on the iterative ADP algorithm is presented which makes the value function converge iteratively to the greatest lower bound of all value function indices within an error according to within finite time. Two neural networks are used as parametric structures to implement the iterative ADP algorithm with -error bound, which aim at approximating the value function and the control policy, respectively. And then, the optimal control policy is obtained. Finally, a simulation example is included to illustrate the applicability of the proposed method.
In this paper, a new online model‐free adaptive dynamic programming algorithm is developed to solve the H∞ control problem of the continuous‐time linear system with completely unknown system dynamics. Solving the game algebraic Riccati equation, commonly used in H∞ state feedback control design, is often referred to as a two‐player differential game where one player tries to minimize the predefined performance index while the other tries to maximize it. Using data generated in real time along the system trajectories, this new method can solve online the game algebraic Riccati equation without requiring the full knowledge of system dynamics. A rigorous proof of convergence of the proposed algorithm is given. Finally, simulation studies on two examples demonstrate the effectiveness of the proposed method.
In this paper, a novel adaptive dynamic programming (ADP)-based event-triggered safe control method is proposed to solve the zero-sum game problem of nonlinear safety-critical systems with safety constraints and input saturation. First, the barrier function-based system transformation, the zerosum game problem with safety constraints and input saturation is transformed into an equivalent input saturation zero-sum game problem, so as to guarantee that the system does not violate the safety constraints. Furthermore, the non-quadratic utility function is introduced into the performance function to solve input saturation. Then, a critic neural network (NN) is constructed to approximate the optimal safety value function. Subsequently, a novel event-triggered scheme is developed to determine the update instant of the control law and the disturbance law. Therefore, the proposed ADP-based event-triggered safe control method can ensure that the states of nonlinear safety-critical systems satisfy the safety constraints, while greatly reducing the amount of calculation and saving communication resources. In addition, during the learning process, the concurrent learning is used to relax the persistence of excitation (PE) condition. According to the Lyapuov theory, it is proved that the weight estimation error of the critic neural network and the states are uniformly ultimately bounded (UUB), and the Zeno behavior is excluded. Finally, a simulation example verifies the effectiveness of the proposed method.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.