We introduce a lightweight and accurate localization method that only utilizes the geometry of 2D-3D lines. Given a pre-captured 3D map, our approach localizes a panorama image, taking advantage of the holistic 360 degree view. The system mitigates potential privacy breaches or domain discrepancies by avoiding trained or hand-crafted visual descriptors. However, as lines alone can be ambiguous, we express distinctive yet compact spatial contexts from relationships between lines, namely the dominant directions of parallel lines and the intersection between non-parallel lines. The resulting representations are efficient in processing time and memory compared to conventional visual descriptor-based methods. Given the groups of dominant line directions and their intersections, we accelerate the search process to test thousands of pose candidates in less than a millisecond without sacrificing accuracy. We empirically show that the proposed 2D-3D matching can localize panoramas for challenging scenes with similar structures, dramatic domain shifts or illumination changes. Our fully geometric approach does not involve extensive parameter tuning or neural network training, making it a practical algorithm that can be readily deployed in the real world.
We consider the task of localizing a panorama image against a 3D line map, which can be readily obtained from images or 3D scans. Unlike existing panoramic localization pipelines, our method performs localization under a fully geometric setup, only using lines in 2D and 3D.
Our method exploits lines and their intersections for performing localization. We first cluster lines using their principal directions in 2D and 3D, which are shown as the line colors in the figure above. Then, we pairwise intersect lines from distinctive princial directions and obtain three groups of intersection points.
Given the initial set of lines and intersections, we perform coarse pose search by comparing point and line distance functions. The distance functions are defined over the sphere as the geodesic distance to the nearest point or line. From uniformly sampled poses in the map, we extract poses that have similar distance function values to those in the query image.
We refine poses by aligning lines and their intersections on the sphere. First, at each pose we project 3D line segments onto the sphere. Then, we perform nearest neighbor matching within each intersection point group and optimize translation by minimizing the spherical distance between the matches. Finally, we refine rotation by aligning the line directions associated with each intersection point match.
1. Junho Kim, Changwoon Choi, Hojun Jang, and Young Min Kim. Ldl: Line distance functions for panoramic localization, ICCV 2023
@InProceedings{Kim_2024_CVPR, author = {Kim, Junho and Jeong, Jiwon and Kim, Young Min}, title = {Fully Geometric Panoramic Localization}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)}, month = {June}, year = {2024}, }