Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/facebookresearch/ov-seg
This is the official PyTorch implementation of the paper Open-Vocabulary Semantic Segmentation with Mask-adapted CLIP.
https://github.com/facebookresearch/ov-seg
Last synced: about 8 hours ago
JSON representation
This is the official PyTorch implementation of the paper Open-Vocabulary Semantic Segmentation with Mask-adapted CLIP.
- Host: GitHub
- URL: https://github.com/facebookresearch/ov-seg
- Owner: facebookresearch
- License: other
- Created: 2022-11-02T05:23:13.000Z (about 2 years ago)
- Default Branch: main
- Last Pushed: 2023-10-17T03:32:27.000Z (about 1 year ago)
- Last Synced: 2024-11-06T12:13:06.442Z (8 days ago)
- Language: Jupyter Notebook
- Size: 16.6 MB
- Stars: 689
- Watchers: 12
- Forks: 61
- Open Issues: 9
-
Metadata Files:
- Readme: README.md
- Contributing: CONTRIBUTING.md
- License: LICENSE
- Code of conduct: CODE_OF_CONDUCT.md
Awesome Lists containing this project
README
# [OVSeg] Open-Vocabulary Semantic Segmentation with Mask-adapted CLIP
This is the official PyTorch implementation of our paper:
**Open-Vocabulary Semantic Segmentation with Mask-adapted CLIP**
[Feng Liang](https://jeff-liangf.github.io/), [Bichen Wu](https://www.linkedin.com/in/bichenwu), [Xiaoliang Dai](https://sites.google.com/view/xiaoliangdai/), [Kunpeng Li](https://kunpengli1994.github.io/), [Yinan Zhao](https://yinan-zhao.github.io/), [Hang Zhang](https://hangzhang.org/), [Peizhao Zhang](https://www.linkedin.com/in/peizhao-zhang-14846042/), [Peter Vajda](https://sites.google.com/site/vajdap), [Diana Marculescu](https://www.ece.utexas.edu/people/faculty/diana-marculescu)
Computer Vision and Pattern Recognition Conference (CVPR), 2023[[arXiv](https://arxiv.org/abs/2210.04150)] [[Project](https://jeff-liangf.github.io/projects/ovseg/)] [[huggingface demo](https://huggingface.co/spaces/facebook/ov-seg)]
## Installation
Please see [installation guide](./INSTALL.md).
## Data Preparation
Please see [datasets preparation](./datasets/DATASETS.md).
## Getting started
Please see [getting started instruction](./GETTING_STARTED.md).
## Finetuning CLIP
Please see [open clip training](./open_clip_training/README.md).
## LICENSE
Shield: [![CC BY-NC 4.0][cc-by-nc-shield]][cc-by-nc]
The majority of OVSeg is licensed under a
[Creative Commons Attribution-NonCommercial 4.0 International License](LICENSE).[![CC BY-NC 4.0][cc-by-nc-image]][cc-by-nc]
[cc-by-nc]: http://creativecommons.org/licenses/by-nc/4.0/
[cc-by-nc-image]: https://licensebuttons.net/l/by-nc/4.0/88x31.png
[cc-by-nc-shield]: https://img.shields.io/badge/License-CC%20BY--NC%204.0-lightgrey.svgHowever portions of the project are under separate license terms: CLIP and ZSSEG are licensed under the [MIT license](https://github.com/openai/CLIP/blob/main/LICENSE); MaskFormer is licensed under the [CC-BY-NC](https://github.com/facebookresearch/MaskFormer/blob/main/LICENSE); openclip is licensed under the license at [its repo](https://github.com/mlfoundations/open_clip/blob/main/LICENSE).
## Citing OVSeg :pray:
If you use OVSeg in your research or wish to refer to the baseline results published in the paper, please use the following BibTeX entry.
```BibTeX
@inproceedings{liang2023open,
title={Open-vocabulary semantic segmentation with mask-adapted clip},
author={Liang, Feng and Wu, Bichen and Dai, Xiaoliang and Li, Kunpeng and Zhao, Yinan and Zhang, Hang and Zhang, Peizhao and Vajda, Peter and Marculescu, Diana},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages={7061--7070},
year={2023}
}
```