https://github.com/VladimirYugay/Gaussian-SLAM
Gaussian-SLAM: Photo-realistic Dense SLAM with Gaussian Splatting
https://github.com/VladimirYugay/Gaussian-SLAM
3d-reconstruction computer-vision gaussian-splatting robotics slam
Last synced: about 1 month ago
JSON representation
Gaussian-SLAM: Photo-realistic Dense SLAM with Gaussian Splatting
- Host: GitHub
- URL: https://github.com/VladimirYugay/Gaussian-SLAM
- Owner: VladimirYugay
- License: mit
- Created: 2023-11-28T09:56:37.000Z (over 1 year ago)
- Default Branch: main
- Last Pushed: 2024-07-03T15:01:05.000Z (10 months ago)
- Last Synced: 2024-08-01T03:33:39.790Z (9 months ago)
- Topics: 3d-reconstruction, computer-vision, gaussian-splatting, robotics, slam
- Language: Python
- Homepage: https://vladimiryugay.github.io/gaussian_slam
- Size: 32.2 MB
- Stars: 840
- Watchers: 56
- Forks: 50
- Open Issues: 3
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
Gaussian-SLAM: Photo-realistic Dense SLAM with Gaussian Splatting
Vladimir Yugay
·
Yue Li*
·
Theo Gevers
·
Martin Oswald
*Significant contribution
Project Page
## ⚙️ Setting Things Up
Clone the repo:
```
git clone https://github.com/VladimirYugay/Gaussian-SLAM
```Make sure that gcc and g++ paths on your system are exported:
```
export CC=
export CXX=
```To find the gcc path and g++ path on your machine you can use which gcc.
Then setup environment from the provided conda environment file,
```
conda env create -f environment.yml
conda activate gslam
```
We tested our code on RTX3090 and RTX A6000 GPUs respectively and Ubuntu22 and CentOS7.5.## 🔨 Running Gaussian-SLAM
Here we elaborate on how to load the necessary data, configure Gaussian-SLAM for your use-case, debug it, and how to reproduce the results mentioned in the paper.
Downloading the Data
We tested our code on Replica, TUM_RGBD, ScanNet, and ScanNet++ datasets. We also provide scripts for downloading Replica nad TUM_RGBD. Install git lfs before using the scripts by running "git lfs install".
For downloading ScanNet, follow the procedure described on here.
For downloading ScanNet++, follow the procedure described on here.
The config files are named after the sequences that we used for our method.
Running the code
Start the system with the command:```
python run_slam.py configs// --input_path --output_path
```
For example:
```
python run_slam.py configs/Replica/room0.yaml --input_path /home/datasets/Replica/room0 --output_path output/Replica/room0
```
You can also configure input and output paths in the config yaml file.
Reproducing Results
While we made all parts of our code deterministic, differential rasterizer of Gaussian Splatting is not. The metrics can be slightly different from run to run. In the paper we report average metrics that were computed over three seeds: 0, 1, and 2.You can reproduce the results for a single scene by running:
```
python run_slam.py configs// --input_path --output_path
```If you are running on a SLURM cluster, you can reproduce the results for all scenes in a dataset by running the script:
```
./scripts/reproduce_sbatch.sh
```
Please note the evaluation of ```depth_L1``` metric requires reconstruction of the mesh, which in turns requires headless installation of open3d if you are running on a cluster.
Demo
We used the camera path tool in gaussian-splatting-lightning repo to help make the fly-through video based on the reconstructed scenes. We thank its author for the great work.
## 📌 Citation
If you find our paper and code useful, please cite us:
```bib
@misc{yugay2023gaussianslam,
title={Gaussian-SLAM: Photo-realistic Dense SLAM with Gaussian Splatting},
author={Vladimir Yugay and Yue Li and Theo Gevers and Martin R. Oswald},
year={2023},
eprint={2312.10070},
archivePrefix={arXiv},
primaryClass={cs.CV}
}