https://github.com/v2ai/selfsup
Collections of self-supervised methods, based on cvpods.
https://github.com/v2ai/selfsup
barlow-twins byol cifar-10 classification contrastive-learning cvpods det-con eqco imagenet moco-v2 point-contrast scrl self-supervised selfsup simclr simo simsiam swav
Last synced: 5 months ago
JSON representation
Collections of self-supervised methods, based on cvpods.
- Host: GitHub
- URL: https://github.com/v2ai/selfsup
- Owner: V2AI
- License: apache-2.0
- Created: 2020-12-07T06:38:31.000Z (almost 5 years ago)
- Default Branch: master
- Last Pushed: 2021-08-21T05:01:14.000Z (about 4 years ago)
- Last Synced: 2025-05-12T19:14:14.464Z (5 months ago)
- Topics: barlow-twins, byol, cifar-10, classification, contrastive-learning, cvpods, det-con, eqco, imagenet, moco-v2, point-contrast, scrl, self-supervised, selfsup, simclr, simo, simsiam, swav
- Language: Python
- Homepage:
- Size: 141 KB
- Stars: 58
- Watchers: 5
- Forks: 8
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# SelfSup
Collections of self-supervised methods (MoCo series, SimCLR, **SiMo**, BYOL, SimSiam, SwAV, PointContrast, etc.).
## Get Started
### Install cvpods following the instructions.
Install cvpods from https://github.com/Megvii-BaseDetection/cvpods.git .
### Prepare Datasets
```shell
cd cvpods
ln -s /path/to/your/ImageNet datasets/imagenet
```### Train your own models
```
cd /path/to/your/SelfSup/examples/simclr/simclr.res50.scratch.imagenet.224size.256bs.200e
# pre-train
pods_train --num-gpus 8
# convert to weights
python convert.py simclr.res50.scratch.imagenet.224size.256bs.200e/log/model_final.pth weights.pkl
# downstream evaluation
cd /path/to/your/simclr.res50.scratch.imagenet.224size.256bs.200e.lin_cls
pods_train --num-gpus 8 MODEL.WEIGHTS /path/to/your/weights.pkl```
## Model Zoo
### Supervised Classification
#### ImageNet
| Methods | Training Schedule | Top 1 Acc |
| ------- | ------ | ------------------ |
| Res50 | 100e | 76.4 |#### CIFAR 10
| Methods | Training Schedule | Top 1 Acc |
| ------- | ------ | ------------------ |
| Res50 | 200e | 95.4 |#### STL 10
| Methods | Training Schedule | Top 1 Acc |
| ------- | ------ | ------------------ |
| Res50 | 150e | 86.1 |### Self-Supervised Learning - Classification
> All results in the below table are trained using resnet-50 and reported on the ILSVRC2012 dataset.
| Methods | Training Schedule | Batch Size | Our Acc@1 | Official Acc@1 |
| ------- | ------ | ---------- | --------- | -------------- |
| MoCo | 200e | 256 | 60.5 | 60.5 |
| MoCov2 | 200e | 256 | **67.6** | 67.5 |
| SimCLR | 200e | 256 | **63.2** | 61.9 |
| **SimCLR*** | 200e | 256 | **67.3** | **Ours** |
| **SiMo** | 200e | 256 | **68.1** | **Ours** |
| SimSiam | 100e | 256 | 67.6 | 67.7 |
| SwAV | 200e | 256 | **73.0** | 72.7 |
| BYOL | 100e | 2048 | **69.8** | 66.5 (bs4096 from SimSiam paper) |
| BarlowTwins | 300e | 1024 | Comming Soon| 71.7 |### Self-Supervised Learning - Detection (2D)
> All the results reported below are trained on ILSVRC2012 and evaluated on MS COCO using Faster-RCNN-FPN and resnet-50.
| Methods | Training Schedule | Batch Size | Box AP |
| ------- | ------ | ---------- | ------------------ |
| SCRL | 200 | 4096 | 39.9 ( official: 40.5 with bs 8192) |
| DetCon | 200 | 256 | Comming Soon. |### Self-Supervised Learning - 3D Scene Understanding
| Methods | Training Schedule | Downstream task |
| ------------- | ----- | --------------- |
| PointContrast | - | Comming Soon. |## Citation
SelfSup is a part of [cvpods](https://github.com/Megvii-BaseDetection/cvpods), so if you find this repo useful in your research, or if you want to refer the implementations in this repo, please consider cite:
```BibTeX
@article{zhu2020eqco,
title={EqCo: Equivalent Rules for Self-supervised Contrastive Learning},
author={Zhu, Benjin and Huang, Junqiang and Li, Zeming and Zhang, Xiangyu and Sun, Jian},
journal={arXiv preprint arXiv:2010.01929},
year={2020}
}@misc{zhu2020cvpods,
title={cvpods: All-in-one Toolbox for Computer Vision Research},
author={Zhu*, Benjin and Wang*, Feng and Wang, Jianfeng and Yang, Siwei and Chen, Jianhu and Li, Zeming},
year={2020}
}
```