Discrete-Time Nonlinear Optimal Control Using Multi-Step Reinforcement Learning

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS(2024)

引用 0|浏览3
暂无评分
摘要
This brief solves the optimal control problem of discrete-time nonlinear systems by proposing a multi-step reinforcement learning (RL) algorithm. The proposed multi-step RL algorithm is established based on the discrete-time optimal Bellman equation, which takes advantage of policy iteration (PI) and value iteration (VI). Benefiting from the multi-step integration mechanism, the algorithm is accelerated. The convergence of multi-step RL is proved by mathematical induction. For real-world implementation purposes, neural network (NN) and Actor-Critic architecture are introduced to approximate the iterative value functions and control policies. A numerical simulation of Chua's circuit illustrates the effectiveness of the proposed algorithm.
更多
查看译文
关键词
Convergence,Optimal control,Heuristic algorithms,Mathematical models,Approximation algorithms,Reinforcement learning,Nonlinear systems,optimal bellman equation,actor-critic architecture,optimal control
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要