Composing Meta-Policies for Autonomous Driving Using Hierarchical Deep Reinforcement Learning.

arXiv: Artificial Intelligence(2017)

引用 26|浏览114
暂无评分
摘要
Rather than learning new control policies for each new task, it is possible, when tasks share some structure, to compose a from previously learned policies. This paper reports results from experiments using Deep Reinforcement Learning on a continuous-state, discrete-action autonomous driving simulator. We explore how Deep Neural Networks can represent meta-policies that switch among a set of previously learned policies, specifically in settings where the dynamics of a new scenario are composed of a mixture of previously learned dynamics and where the state observation is possibly corrupted by sensing noise. We also report the results of experiments varying dynamics mixes, distractor policies, magnitudes/distributions of sensing noise, and obstacles. In a fully observed experiment, the meta-policy learning algorithm achieves 2.6x the reward achieved by the next best policy composition technique with 80% less exploration. In a partially observed experiment, the meta-policy learning algorithm converges after 50 iterations while a direct application of RL fails to converge even after 200 iterations.
更多
查看译文
关键词
hierarchical deep reinforcement learning,deep reinforcement learning,autonomous driving,meta-policies
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要