Abstract
We introduce RGS-SLAM, a robust Gaussian-splatting SLAM framework that replaces the residual-driven densification stage of GS-SLAM with a training-free correspondence-to-Gaussian initialization. Instead of progressively adding Gaussians as residuals reveal missing geometry, RGS-SLAM performs a one-shot triangulation of dense multi-view correspondences derived from DINOv3 descriptors refined through a confidence-aware inlier classifier, generating a well-distributed and structure-aware Gaussian seed prior to optimization. This initialization stabilizes early mapping and accelerates convergence by roughly 20%, yielding higher rendering fidelity in texture-rich and cluttered scenes while remaining fully compatible with existing GS-SLAM pipelines. Evaluated on the TUM RGB-D and Replica datasets, RGS-SLAM achieves competitive or superior localization and reconstruction accuracy compared with state-of-the-art Gaussian and point-based SLAM systems, sustaining real-time mapping performance at up to 925 FPS.
Rendering results on the Replica dataset.
Rendering results on the TUM RGB-D dataset.
Tracking Results
Result on Replica dataset
Result on TUM RGB-D dataset
BibTeX
@article{rgs-slam-2026,
title = {RGS-SLAM: Robust Gaussian Splatting SLAM with One-Shot Dense Initialization},
author = {Wei-Tse Cheng, Yen-Jen Chiou, Yuan-Fu Yang},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
year = {2026},
url = {https://github.com/Breeze1124/RGS-SLAM}
}