Localization Is All You Evaluate: Data Leakage in Online Mapping Datasets and How to Fix It
Paper in proceeding, 2024
State-of-the-art methods are based on supervised learning and are trained predominantly using two datasets: nuScenes and Argoverse 2.
However, these datasets revisit the same geographic locations across training, validation, and test sets.
Specifically, over 80% of nuScenes and 40% of Argoverse 2 validation and test samples are less than 5 m from a training sample.
At test time, the methods are thus evaluated more on how well they localize within a memorized implicit map built from the training data than on extrapolating to unseen locations.
Naturally, this data leakage causes inflated performance numbers and we propose geographically disjoint data splits to reveal the true performance in unseen environments.
Experimental results show that methods perform considerably worse, some dropping more than 45 mAP, when trained and evaluated on proper data splits.
Additionally, a reassessment of prior design choices reveals diverging conclusions from those based on the original split.
Notably, the impact of lifting methods and the support from auxiliary tasks (e.g., depth supervision) on performance appears less substantial or follows a different trajectory than previously perceived.
Computer Vision
Autonomous Driving
Online Mapping
Author
Adam Lilja
Chalmers, Electrical Engineering, Signal Processing and Biomedical Engineering
Junsheng Fu
Erik Stenborg
Chalmers, Electrical Engineering, Signal Processing and Biomedical Engineering
Lars Hammarstrand
Chalmers, Electrical Engineering, Signal Processing and Biomedical Engineering
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
10636919 (ISSN)
Seattle, USA,
Subject Categories
Computer Vision and Robotics (Autonomous Systems)
DOI
10.48550/arXiv.2312.06420