QC_SANE: Robust Control in DRL Using Quantile Critic With Spiking Actor and Normalized Ensemble

IEEE transactions on neural networks and learning systems(2023)

引用 2|浏览5
暂无评分
摘要
Recently introduced deep reinforcement learning (DRL) techniques in discrete-time have resulted in significant advances in online games, robotics, and so on. Inspired from recent developments, we have proposed an approach referred to as Quantile Critic with Spiking Actor and Normalized Ensemble (QC_SANE) for continuous control problems, which uses quantile loss to train critic and a spiking neural network (NN) to train an ensemble of actors. The NN does an internal normalization using a scaled exponential linear unit (SELU) activation function and ensures robustness. The empirical study on multijoint dynamics with contact (MuJoCo)-based environments shows improved training and test results than the state-of-the-art approach: population coded spiking actor network (PopSAN).
更多
查看译文
关键词
Artificial neural networks,Neurons,Uncertainty,Task analysis,Robustness,Statistics,Sociology,Actor critic,deep reinforcement learning (DRL),ensemble,reinforcement learning (RL),robust control,spiking neural network (SNN)
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要