Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/ldkong1205/LaserMix
[CVPR 2023 Highlight] LaserMix for Semi-Supervised LiDAR Semantic Segmentation
https://github.com/ldkong1205/LaserMix
autonomous-driving lidar segmentation semi-supervised-learning
Last synced: 12 days ago
JSON representation
[CVPR 2023 Highlight] LaserMix for Semi-Supervised LiDAR Semantic Segmentation
- Host: GitHub
- URL: https://github.com/ldkong1205/LaserMix
- Owner: ldkong1205
- License: apache-2.0
- Created: 2022-06-21T08:59:57.000Z (over 2 years ago)
- Default Branch: main
- Last Pushed: 2024-05-14T03:39:28.000Z (6 months ago)
- Last Synced: 2024-08-01T03:42:41.408Z (3 months ago)
- Topics: autonomous-driving, lidar, segmentation, semi-supervised-learning
- Language: Python
- Homepage: https://ldkong.com/LaserMix
- Size: 8.61 MB
- Stars: 279
- Watchers: 16
- Forks: 17
- Open Issues: 3
-
Metadata Files:
- Readme: README.md
- License: LICENSE
- Citation: CITATION.cff
Awesome Lists containing this project
- Awesome-Mixup - [Code
README
LaserMix for Semi-Supervised LiDAR Semantic Segmentation
Lingdong Kong,
Jiawei Ren,
Liang Pan,
Ziwei Liu
S-Lab, Nanyang Technological University
## About
LaserMix is a semi-supervised learning (SSL) framework designed for LiDAR semantic segmentation. It leverages the strong spatial prior of driving scenes to construct low-variation areas via laser beam mixing, and encourages segmentation models to make confident and consistent predictions before and after mixing.
Fig. Illustration for laser beam partition based on inclination φ.
Visit our project page to explore more details. :red_car:
## Updates
- \[2024.05\] - Our improved framework, [LaserMix++](https://arxiv.org/abs/2405.05258) :rocket:, is avaliable on arXiv.
- \[2024.01\] - The [toolkit](https://github.com/robodrive-24/toolkit) tailored for [The RoboDrive Challenge](https://robodrive-24.github.io/) has been released. :hammer_and_wrench:
- \[2023.12\] - We are hosting [The RoboDrive Challenge](https://robodrive-24.github.io/) at [ICRA 2024](https://2024.ieee-icra.org/). :blue_car:
- \[2023.12\] - Introducing [FRNet](https://github.com/Xiangxu-0103/FRNet), an efficient and effective **real-time** LiDAR segmentation model that achieves promising semi-supervised learning results on `SemanticKITTI` and `nuScenes`. Code and checkpoints are available for downloading.
- \[2023.03\] - Intend to test the **robustness** of your LiDAR semantic segmentation models? Check our recent work, :robot: [Robo3D](https://github.com/ldkong1205/Robo3D), a comprehensive suite that enables OoD robustness evaluation of 3D segmentors on our newly established datasets: `SemanticKITTI-C`, `nuScenes-C`, and `WOD-C`.
- \[2023.03\] - LaserMix was selected as a :sparkles: highlight :sparkles: at [CVPR 2023](https://cvpr.thecvf.com/) (top 10% of accepted papers).
- \[2023.02\] - LaserMix was accepted to [CVPR 2023](https://cvpr.thecvf.com/)! :tada:
- \[2023.02\] - LaserMix has been integrated into the [MMDetection3D](https://github.com/open-mmlab/mmdetection3d) codebase! Check [this](https://github.com/open-mmlab/mmdetection3d/pull/2302) PR in the `dev-1.x` branch to know more details. :beers:
- \[2023.01\] - As suggested, we will establish a *sequential track* taking into account the LiDAR data collection nature in our semi-supervised LiDAR semantic segmentation benchmark. The results will be gradually updated in [RESULT.md](docs/RESULT.md).
- \[2022.12\] - We support a wider range of LiDAR segmentation backbones, including [RangeNet++](https://www.ipb.uni-bonn.de/wp-content/papercite-data/pdf/milioto2019iros.pdf), [SalsaNext](https://arxiv.org/abs/2003.03653), [FIDNet](https://arxiv.org/abs/2109.03787), [CENet](https://arxiv.org/abs/2207.12691), [MinkowskiUNet](https://github.com/NVIDIA/MinkowskiEngine), [Cylinder3D](https://openaccess.thecvf.com/content/CVPR2021/papers/Zhu_Cylindrical_and_Asymmetrical_3D_Convolution_Networks_for_LiDAR_Segmentation_CVPR_2021_paper.pdf), and [SPVCNN](https://arxiv.org/pdf/2007.16100), under both fully- and semi-supervised settings. The checkpoints will be available soon!
- \[2022.12\] - The derivation of spatial-prior-based SSL is available [here](https://ldkong.com/LaserMix/derivation.pdf). Take a look! :memo:
- \[2022.08\] - LaserMix achieves 1st place among the semi-supervised semantic segmentation leaderboards of [nuScenes](https://paperswithcode.com/sota/semi-supervised-semantic-segmentation-on-25), [SemanticKITTI](https://paperswithcode.com/sota/semi-supervised-semantic-segmentation-on-24), and [ScribbleKITTI](https://paperswithcode.com/sota/semi-supervised-semantic-segmentation-on-23), based on [Paper-with-Code](https://paperswithcode.com/paper/lasermix-for-semi-supervised-lidar-semantic). :bar_chart:
- \[2022.08\] - We provide a [video demo](https://youtu.be/Xkwa5-dT0g4) for visual comparisons on the SemanticKITTI val set. Take a look!
- \[2022.07\] - Our paper is available on arXiv, click here to check it out. Code will be available soon!## Outline
- [Installation](#installation)
- [Data Preparation](#data-preparation)
- [Getting Started](#getting-started)
- [Video Demo](#video-demo)
- [Main Results](#main-results)
- [TODO List](#todo-list)
- [License](#license)
- [Acknowledgement](#acknowledgement)
- [Citation](#citation)## Installation
Please refer to [INSTALL.md](docs/INSTALL.md) for the installation details.## Data Preparation
Please refer to [DATA_PREPARE.md](docs/DATA_PREPARE.md) for the details to prepare the 1[nuScenes](https://www.nuscenes.org), 2[SemanticKITTI](http://www.semantic-kitti.org/), and 3[ScribbleKITTI](https://github.com/ouenal/scribblekitti) datasets.## Getting Started
Please refer to [GET_STARTED.md](docs/GET_STARTED.md) to learn more usage about this codebase.## Video Demo
| Demo 1 | Demo 2| Demo 3 |
| :-: | :-: | :-: |
| | | |
| [Link](https://youtu.be/Xkwa5-dT0g4) :arrow_heading_up: | [Link](https://youtu.be/OlKNDt8_um4) :arrow_heading_up: | [Link](https://youtu.be/f8UKgxi5mow) :arrow_heading_up: |## Main Result
### Framework Overview
### Range View
Method
nuScenes
SemanticKITTI
ScribbleKITTI
1% 10% 20% 50%
1% 10% 20% 50%
1% 10% 20% 50%
Sup.-only
38.3 57.5 62.7 67.6
36.2 52.2 55.9 57.2
33.1 47.7 49.9 52.5
LaserMix
49.568.270.673.0
43.458.859.461.4
38.354.455.658.7
improv. ↑
+11.2 +10.7 +7.9 +5.4
+7.2 +6.6 +3.5 +4.2
+5.2 +6.7 +5.7 +6.2
LaserMix++
improv. ↑
### Voxel
Method
nuScenes
SemanticKITTI
ScribbleKITTI
1% 10% 20% 50%
1% 10% 20% 50%
1% 10% 20% 50%
Sup.-only
50.9 65.9 66.6 71.2
45.4 56.1 57.8 58.7
39.2 48.0 52.1 53.8
LaserMix
55.3 69.9 71.8 73.2
50.6 60.0 61.9 62.3
44.2 53.7 55.1 56.8
improv. ↑
+4.4 +4.0 +5.2 +2.0
+5.2 +3.9 +4.1 +3.6
+5.0 +5.7 +3.0 +3.0
LaserMix++
improv. ↑
### Ablation Studies
### Qualitative Examples
![qualitative](docs/figs/qualitative.png)
### Checkpoints & More Results
For more experimental results and pretrained weights, please refer to [RESULT.md](docs/RESULT.md).
## TODO List
- [x] Initial release. :rocket:
- [x] Add license. See [here](#license) for more details.
- [x] Add video demos :movie_camera:
- [x] Add installation details.
- [x] Add data preparation details.
- [ ] Add evaluation details.
- [ ] Add training details.## Citation
If you find this work helpful, please kindly consider citing our paper:
```bibtex
@inproceedings{kong2023lasermix,
title = {LaserMix for Semi-Supervised LiDAR Semantic Segmentation},
author = {Kong, Lingdong and Ren, Jiawei and Pan, Liang and Liu, Ziwei},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages = {21705--21715},
year = {2023},
}
```## License
This work is under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.## Acknowledgement
This work is developed based on the [MMDetection3D](https://github.com/open-mmlab/mmdetection3d) codebase.>
> MMDetection3D is an open-source toolbox based on PyTorch, towards the next-generation platform for general 3D perception. It is a part of the OpenMMLab project developed by MMLab.We acknowledge the use of the following public resources during the course of this work: 1[nuScenes](https://www.nuscenes.org), 2[nuScenes-devkit](https://github.com/nutonomy/nuscenes-devkit), 3[SemanticKITTI](http://www.semantic-kitti.org/), 4[SemanticKITTI-API](https://github.com/PRBonn/semantic-kitti-api), 5[ScribbleKITTI](https://github.com/ouenal/scribblekitti), 6[FIDNet](https://github.com/placeforyiming/IROS21-FIDNet-SemanticKITTI), 7[CENet](https://github.com/huixiancheng/CENet), 8[SPVNAS](https://github.com/mit-han-lab/spvnas), 9[Cylinder3D](https://github.com/xinge008/Cylinder3D), 10[TorchSemiSeg](https://github.com/charlesCXK/TorchSemiSeg), 11[MixUp](https://github.com/facebookresearch/mixup-cifar10), 12[CutMix](https://github.com/clovaai/CutMix-PyTorch), 13[CutMix-Seg](https://github.com/Britefury/cutmix-semisup-seg), 14[CBST](https://github.com/yzou2/CBST), 15[MeanTeacher](https://github.com/CuriousAI/mean-teacher), and 16[Cityscapes](https://www.cityscapes-dataset.com).
We would like to thank Fangzhou Hong for the insightful discussions and feedback. ❤️