Camera-Radar Fusion for 3-D Depth Reconstruction

2020 IEEE Intelligent Vehicles Symposium (IV)(2020)

引用 2|浏览66
暂无评分
摘要
We introduce and study the problem of camera-radar fusion for 3-D depth reconstruction. This problem is motivated by autonomous driving applications, in which we can expect to have access to both front-facing camera and radar sensors. These two sensors are complementary in several respects: the camera is a passive sensor measuring azimuth and elevation; the radar is an active sensor measuring azimuth and range. Fusing their measurements is therefore beneficial. Our fusion solution uses a modified encoder-decoder deep convolutional neural network. We train and evaluate this network on over 100 000 samples collected in highway environments. Our results demonstrate an improvement in reconstruction accuracy and robustness from fusing the two sensors.
更多
查看译文
关键词
autonomous driving applications,radar sensors,passive sensor,active sensor measuring azimuth,fusion solution,modified encoder-decoder deep convolutional neural network,camera-radar fusion,3-D depth reconstruction
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要