The LaMAria Dataset and Benchmark
Benchmarking Egocentric Visual-Inertial SLAM at City Scale
Abstract
Precise 6-DoF simultaneous localization and mapping (SLAM) from onboard sensors is critical for wearable devices capturing egocentric data, which exhibits specific challenges, such as a wider diversity of motions and viewpoints, prevalent dynamic visual content, or long sessions affected by time-varying sensor calibration. While recent progress on SLAM has been swift, academic research is still driven by benchmarks that do not reflect these challenges or do not offer sufficiently accurate ground truth poses. In this paper, we introduce a new dataset and benchmark for visual-inertial SLAM with egocentric, multi-modal data. We record hours and kilometers of trajectories through a city center with glasses-like devices equipped with various sensors. We leverage surveying tools to obtain control points as indirect pose annotations that are metric, centimeter-accurate, and available at city scale. This makes it possible to evaluate extreme trajectories that involve walking at night or traveling in a vehicle. We show that state-of-the-art systems developed by academia are not robust to these challenges and we identify components that are responsible for this. In addition, we design tracks with different levels of difficulty to ease in-depth analysis and evaluation of less mature approaches.
Overview


BibTeX
@InProceedings{Krishnan_2025_ICCV, author = {Krishnan, Anusha and Liu, Shaohui and Sarlin, Paul-Edouard and Gentilhomme, Oscar and Caruso, David and Monge, Maurizio and Newcombe, Richard and Engel, Jakob and Pollefeys, Marc}, title = {Benchmarking Egocentric Visual-Inertial SLAM at City Scale}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, year = {2025} }