Localization is a critically essential and crucial enabler of autonomous robots. While deep learning has made significant strides in many computer vision tasks, it is still yet to make a sizeable impact on improving capabilities of metric visual localization. One of the major hindrances has been the inability of existing Convolutional Neural Network (CNN)- based pose regression methods to generalize to previously unseen places. Our recently introduced CMRNet effectively addresses this limitation by enabling map independent monocular localization in LiDAR-maps. In this paper, we now take it a step further by introducing CMRNet++, which is a significantly more robust model that not only generalizes to new places effectively, but is also independent of the camera parameters. We enable this capability by combining deep learning with geometric techniques, and by moving the metric reasoning outside the learning process. In this way, the weights of the network are not tied to a specific camera. Extensive evaluations of CMRNet++ on three challenging autonomous driving datasets, i.e., KITTI, Argoverse, and Lyft5, show that CMRNet++ outperforms CMRNet as well as other baselines by a large margin. More importantly, for the first-time, we demonstrate the ability of a deep learning approach to accurately localize without any retraining or fine-tuning in a completely new environment and independent of the camera parameters.

Sorrenti, D., Cattaneo, D., Abhinav, V. (2020). CMRNet++: Map and camera agnostic monocular visual localization in lidar maps. In Proceeding of ICRA 2020 Workshop on Emerging Learning and Algorithmic Methods for Data Association in Robotics https://sites.google.com/view/edat/home also available as arxiv report https://arxiv.org/abs/2004.13795.

CMRNet++: Map and camera agnostic monocular visual localization in lidar maps

Sorrenti;
2020

Abstract

Localization is a critically essential and crucial enabler of autonomous robots. While deep learning has made significant strides in many computer vision tasks, it is still yet to make a sizeable impact on improving capabilities of metric visual localization. One of the major hindrances has been the inability of existing Convolutional Neural Network (CNN)- based pose regression methods to generalize to previously unseen places. Our recently introduced CMRNet effectively addresses this limitation by enabling map independent monocular localization in LiDAR-maps. In this paper, we now take it a step further by introducing CMRNet++, which is a significantly more robust model that not only generalizes to new places effectively, but is also independent of the camera parameters. We enable this capability by combining deep learning with geometric techniques, and by moving the metric reasoning outside the learning process. In this way, the weights of the network are not tied to a specific camera. Extensive evaluations of CMRNet++ on three challenging autonomous driving datasets, i.e., KITTI, Argoverse, and Lyft5, show that CMRNet++ outperforms CMRNet as well as other baselines by a large margin. More importantly, for the first-time, we demonstrate the ability of a deep learning approach to accurately localize without any retraining or fine-tuning in a completely new environment and independent of the camera parameters.
slide + paper
deep learning, robotics, localization, sensor fusion
English
ICRA 2020 Workshop on Emerging Learning and Algorithmic Methods for Data Association in Robotics
2020
Proceeding of ICRA 2020 Workshop on Emerging Learning and Algorithmic Methods for Data Association in Robotics https://sites.google.com/view/edat/home also available as arxiv report https://arxiv.org/abs/2004.13795
2020
none
Sorrenti, D., Cattaneo, D., Abhinav, V. (2020). CMRNet++: Map and camera agnostic monocular visual localization in lidar maps. In Proceeding of ICRA 2020 Workshop on Emerging Learning and Algorithmic Methods for Data Association in Robotics https://sites.google.com/view/edat/home also available as arxiv report https://arxiv.org/abs/2004.13795.
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/328588
Citazioni
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
Social impact