TY - GEN
T1 - Sensor modality fusion with CNNs for UGV autonomous driving in indoor environments
AU - Patel, Naman
AU - Choromanska, Anna
AU - Krishnamurthy, Prashanth
AU - Khorrami, Farshad
PY - 2017/12/13
Y1 - 2017/12/13
N2 - We present a novel end-to-end learning framework to enable ground vehicles to autonomously navigate unknown environments by fusing raw pixels from cameras and depth measurements from a LiDAR. A deep neural network architecture is introduced to effectively perform modality fusion and reliably predict steering commands even in the presence of sensor failures. The proposed network is trained on our own dataset, from LiDAR and a camera mounted on a UGV taken in an indoor corridor environment. Comprehensive experimental evaluation to demonstrate the robustness of our network architecture is performed to show that the proposed deep learning neural network is able to autonomously navigate in the corridor environment. Furthermore, we demonstrate that the fusion of the camera and LiDAR modalities provides further benefits beyond robustness to sensor failures. Specifically, the multimodal fused system shows a potential to navigate around static and dynamic obstacles and to handle changes in environment geometry without being trained for these tasks.
AB - We present a novel end-to-end learning framework to enable ground vehicles to autonomously navigate unknown environments by fusing raw pixels from cameras and depth measurements from a LiDAR. A deep neural network architecture is introduced to effectively perform modality fusion and reliably predict steering commands even in the presence of sensor failures. The proposed network is trained on our own dataset, from LiDAR and a camera mounted on a UGV taken in an indoor corridor environment. Comprehensive experimental evaluation to demonstrate the robustness of our network architecture is performed to show that the proposed deep learning neural network is able to autonomously navigate in the corridor environment. Furthermore, we demonstrate that the fusion of the camera and LiDAR modalities provides further benefits beyond robustness to sensor failures. Specifically, the multimodal fused system shows a potential to navigate around static and dynamic obstacles and to handle changes in environment geometry without being trained for these tasks.
UR - http://www.scopus.com/inward/record.url?scp=85041958768&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85041958768&partnerID=8YFLogxK
U2 - 10.1109/IROS.2017.8205958
DO - 10.1109/IROS.2017.8205958
M3 - Conference contribution
AN - SCOPUS:85041958768
T3 - IEEE International Conference on Intelligent Robots and Systems
SP - 1531
EP - 1536
BT - IROS 2017 - IEEE/RSJ International Conference on Intelligent Robots and Systems
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2017
Y2 - 24 September 2017 through 28 September 2017
ER -