SeqMatchNet: Contrastive Learning with Sequence Matching for Place Recognition and Relocalization

, Vankadari, Madhu, & (2021) SeqMatchNet: Contrastive Learning with Sequence Matching for Place Recognition and Relocalization. In Faust, Aleksandra, Hsu, David, & Neumann, Gerhard (Eds.) Proceedings of the 5th Conference on Robot Learning (CoRL). Proceedings of Machine Learning Research, United States of America, pp. 429-443.

Free-to-read version at publisher website

Description

Visual Place Recognition (VPR) for mobile robot global relocalization is a well-studied problem, where contrastive learning based representation training methods have led to state-of-the-art performance. However, these methods are mainly designed for single image based VPR, where sequential information, which is ubiquitous in robotics, is only used as a post-processing step for filtering single image match scores, but is never used to guide the representation learning process itself. In this work, for the first time, we bridge the gap between single image representation learning and sequence matching through SeqMatchNet which transforms the single image descriptors such that they become more responsive to the sequence matching metric. We propose a novel triplet loss formulation where the distance metric is based on sequence matching, that is, the aggregation of temporal order-based Euclidean distances computed using single images. We use the same metric for mining negatives online during the training which helps the optimization process by selecting appropriate positives and harder negatives. To overcome the computational overhead of sequence matching for negative mining, we propose a 2D convolution based formulation of sequence matching for efficiently aggregating distances within a distance matrix computed using single images. We show that our proposed method achieves consistent gains in performance as demonstrated on four benchmark datasets. Source code available at https://github.com/oravus/SeqMatchNet.

Impact and interest:

14 citations in Scopus
Search Google Scholar™

Citation counts are sourced monthly from Scopus and Web of Science® citation databases.

These databases contain citations from different subsets of available publications and different time periods and thus the citation count from each is usually different. Some works are not in either database and no count is displayed. Scopus includes citations from articles published in 1996 onwards, and Web of Science® generally from 1980 onwards.

Citations counts from the Google Scholar™ indexing service can be viewed at the linked Google Scholar™ search.

ID Code: 243381
Item Type: Chapter in Book, Report or Conference volume (Conference contribution)
Series Name: Proceedings of Machine Learning Research
ORCID iD:
Garg, Souravorcid.org/0000-0001-6068-3307
Milford, Michaelorcid.org/0000-0002-5162-1793
Additional Information: Acknowledgments: We acknowledge the continued support from the QUT Centre for Robotics and thank the reviewers and the editor for their useful comments. We used Weights & Biases [104] for experiment tracking for this paper.
Measurements or Duration: 15 pages
Keywords: Contrastive Learning, Localization, Visual Place Recognition
Pure ID: 146705903
Divisions: Current > Research Centres > Centre for Robotics
Current > Research Centres > Centre for Future Mobility/CARRSQ
Current > QUT Faculties and Divisions > Faculty of Engineering
Current > Schools > School of Electrical Engineering & Robotics
Current > QUT Faculties and Divisions > Faculty of Health
Funding Information: We acknowledge the continued support from the QUT Centre for Robotics and thank the reviewers and the editor for their useful comments. We used Weights & Biases [104] for experiment tracking for this paper.
Copyright Owner: Consult author(s) regarding copyright matters
Copyright Statement: This work is covered by copyright. Unless the document is being made available under a Creative Commons Licence, you must assume that re-use is limited to personal use and that permission from the copyright owner must be obtained for all other uses. If the document is available under a Creative Commons License (or other specified license) then refer to the Licence for details of permitted re-use. It is a condition of access that users recognise and abide by the legal requirements associated with these rights. If you believe that this work infringes copyright please provide details by email to qut.copyright@qut.edu.au
Deposited On: 04 Oct 2023 03:00
Last Modified: 02 Aug 2024 02:58