Integrating State-Of-The-Art Cnns For Multi-Sensor 3d Vehicle Detection In Real Autonomous Driving Environments

2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC)(2019)

引用 8|浏览40
暂无评分
摘要
This paper presents two new approaches to detect surrounding vehicles in 3D urban driving scenes and their corresponding Bird's Eye View (BEV). The proposals integrate two state-of-the-art Convolutional Neural Networks (CNNs), such as YOLOv3 and Mask-RCNN, in a framework presented by the authors in [1] for 3D vehicles detection fusing semantic image segmentation and LIDAR point cloud. Our proposals take advantage of multimodal fusion, geometrical constrains, and pre-trained modules inside our framework. The methods have been tested using the KITTI object detection benchmark and comparison is presented. Experiments show new approaches improve results with respect to the baseline and are on par with other competitive state-of-the-art proposals, being the only ones that do not apply an end-to-end learning process. In this way, they remove the need to train on a specific dataset and show a good capability of generalization to any domain, a key point for self-driving systems. Finally, we have tested our best proposal in KITTI in our driving environment, without any adaptation, obtaining results suitable for our autonomous driving application.
更多
查看译文
关键词
end-to-end learning process,autonomous driving application,multisensor 3D vehicle detection,autonomous driving environments,3D urban driving scenes,YOLOv3,Mask-RCNN,semantic image segmentation,LIDAR point cloud,multimodal fusion,Convolutional Neural Networks,Birds Eye View
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要