SALSA: Swift Adaptive Lightweight Self-Attention for Enhanced LiDAR Place Recognition

Raktim Gautam Goswami, Naman Patel, Prashanth Krishnamurthy, Farshad Khorrami

Research output: Contribution to journalArticlepeer-review

Abstract

Large-scale LiDAR mappings and localization leverage place recognition techniques to mitigate odometry drifts, ensuring accurate mapping. These techniques utilize scene representations from LiDAR point clouds to identify previously visited sites within a database. Local descriptors, assigned to each point within a point cloud, are aggregated to form a scene representation for the point cloud. These descriptors are also used to re-rank the retrieved point clouds based on geometric fitness scores. We propose SALSA, a novel, lightweight, and efficient framework for LiDAR place recognition. It consists of a Sphereformer backbone that uses radial window attention to enable information aggregation for sparse distant points, an adaptive self-attention layer to pool local descriptors into tokens, and a multi-layer-perceptron Mixer layer for aggregating the tokens to generate a scene descriptor. The proposed framework outperforms existing methods on various LiDAR place recognition datasets in terms of both retrieval and metric localization while operating in real-time.

Original languageEnglish (US)
Pages (from-to)8242-8249
Number of pages8
JournalIEEE Robotics and Automation Letters
Volume9
Issue number10
DOIs
StatePublished - 2024

Keywords

  • Deep learning for visual perception
  • deep learning methods
  • localization
  • representation learning

ASJC Scopus subject areas

  • Control and Systems Engineering
  • Biomedical Engineering
  • Human-Computer Interaction
  • Mechanical Engineering
  • Computer Vision and Pattern Recognition
  • Computer Science Applications
  • Control and Optimization
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'SALSA: Swift Adaptive Lightweight Self-Attention for Enhanced LiDAR Place Recognition'. Together they form a unique fingerprint.

Cite this