Awesome-Anything
General AI methods for Anything: AnyObject, AnyGeneration, AnyModel, AnyTask, AnyX
https://github.com/VainF/Awesome-Anything
Last synced: 5 days ago
JSON representation
-
AnyObject
-  <br> *Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alex Berg, Wan-Yen Lo, Piotr Dollar, Ross Girshick* <br> > Meta Research <br> > Preprint'23 <br><br> [[**Segment Anything (Project)**](https://github.com/facebookresearch/segment-anything)] |  | [[Github](https://github.com/facebookresearch/segment-anything)] <br> [[Page](https://segment-anything.com/)] <br> [[Demo](https://segment-anything.com/demo)] |
-  <br> *Feng Liang, Bichen Wu, Xiaoliang Dai, Kunpeng Li, Yinan Zhao, Hang Zhang, Peizhao Zhang, Peter Vajda, Diana Marculescu* <br> > Meta Research <br> > Preprint'23 <br><br> [[**OVSeg (Project)**](https://github.com/facebookresearch/segment-anything)] | <img width="855" alt="image" src="https://user-images.githubusercontent.com/18592211/232279307-cf00ebe2-0751-48dc-b4ac-47ff343c28dc.png"> | [[Github](https://github.com/facebookresearch/ov-seg)] <br> [[Page](https://jeff-liangf.github.io/projects/ovseg/)] |
- ] <br> [[Page](https://github.com/ronghanghu/seg_every_thing)] |
-  <br> *Shilong Liu and Zhaoyang Zeng and Tianhe Ren and Feng Li and Hao Zhang and Jie Yang and Chunyuan Li and Jianwei Yang and Hang Su and Jun Zhu and Lei Zhang* <br> > IDEA-Research <br> > Preprint'23 <br><br> [[**Grounded-SAM**](https://github.com/IDEA-Research/Grounded-Segment-Anything), [**GroundingDINO (Project)**](https://github.com/IDEA-Research/GroundingDINO)] |  | [[Github](https://github.com/IDEA-Research/Grounded-Segment-Anything)] <br> [[Demo](https://colab.research.google.com/github/roboflow-ai/notebooks/blob/main/notebooks/zero-shot-object-detection-with-grounding-dino.ipynb)] |
- ] | <img width="903" alt="image" src="https://user-images.githubusercontent.com/18592211/230897227-c797f375-a44d-4536-a06b-41f0d9f4dbc4.png"> | [[Github](https://github.com/baaivision/Painter)] |
-  <br> Kadir Nar |  | [[Github](https://github.com/kadirnar/segment-anything-video)] |
-  <br> *Achal Dave, Pavel Tokmakov, Deva Ramanan* <br> > ICCV'19 Workshop <br><br> [[**segment-any-moving (Project)**](https://github.com/achalddave/segment-any-moving)] | [<img src="http://www.achaldave.com/projects/anything-that-moves/videos/ZXN6A-tracked-with-objectness-trimmed.gif" width="32%" />](http://www.achaldave.com/projects/anything-that-moves/videos/ZXN6A-tracked-with-objectness-trimmed.mp4)[<img src="http://www.achaldave.com/projects/anything-that-moves/videos/c95cd17749.gif" width="32%" />](http://www.achaldave.com/projects/anything-that-moves/videos/c95cd17749.mp4)<img src="http://www.achaldave.com/projects/anything-that-moves/videos/e0bdb5dfae.gif" width="32%" /> | [[Github](https://github.com/achalddave/segment-any-moving)] |
-  <br> *Jiaqi Chen, Zeyu Yang, Li Zhang* <br><br> [[**Semantic-Segment-Anything (Project)**](https://github.com/fudan-zvg/Semantic-Segment-Anything)] | <img width="903" alt="image" src="https://github.com/fudan-zvg/Semantic-Segment-Anything/blob/main/figures/SSA_motivation.png"> | [[Github](https://github.com/fudan-zvg/Semantic-Segment-Anything)] |
-  <br> *Peize Sun* and *Shoufa Chen* |  | [[Github](https://github.com/Cheems-Seminar/segment-anything-and-name-it)]
-  <br> *Yunkang Cao* | <img width="677" alt="image" src="https://user-images.githubusercontent.com/18592211/231068964-ddeae0ea-4e83-40d6-b73e-2811d46f808d.png"> | [[Github](https://github.com/caoyunkang/GroundedSAM-zero-shot-anomaly-detection)] |
- ] |
-  <br> *Rockey* |  | [[Github](https://github.com/RockeyCoss/Prompt-Segment-Anything)]|
-  <br> *Qingyun Li* |  | [[Github](https://github.com/Li-Qingyun/sam-mmrotate)] |
-  <br> *Aliaksandr Hancharenka, Alexander Chichigin* |  | [[Github](https://github.com/aliaksandr960/segment-anything-eo)] |
-  <br> *Jordão Bragantini, Kyle I S Harrington, Ajinkya Kulkarni* | <img width="658" alt="image" src="https://user-images.githubusercontent.com/18592211/231413725-661fb2a9-1951-40b1-8239-6896eeb7eb4c.png"> | [[Github](https://github.com/JoOkuma/napari-segment-anything)] |
-  <br> *Jordão Bragantini, Kyle I S Harrington, Ajinkya Kulkarni* |  | [[Github](https://github.com/amine0110/SAM-Medical-Imaging)] |
-  <br> *Zhenhua Yang, Qing Jiang* |  | [[Github](https://github.com/yeungchenwa/OCR-SAM)] |
-  <br> *MaybeShewill-CV* |  | [[Github](https://github.com/MaybeShewill-CV/segment-anything-u-specify)] |
-  <br> *Xueyan Zou, Jianwei Yang, Hao Zhang, Feng Li, Linjie Li, Jianfeng Gao, Yong Jae Lee* <br><br> [[SEEM (Project)](https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once)] |  | [[Github](https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once)] |
-  <br> *Harry* |  | [[Github](https://github.com/lujiazho/SegDrawer)] |
-  <br> *Harry* | <img width="546" alt="image" src="https://user-images.githubusercontent.com/18592211/232190851-1dc85342-3d50-42a7-a8e2-f45c4c862d70.png"> | [[Github](https://github.com/kevmo314/magic-copy)] |
-  <br> *Jinyu Yang, Mingqi Gao, Zhe Li, Shang Gao, Fangjing Wang, Feng Zheng* <br><br> [[Track-Anything (Project)](https://github.com/gaomingqi/Track-Anything)] |  | [[Github](https://github.com/gaomingqi/Track-Anything)] <br> [[Demo](https://huggingface.co/spaces/watchtowerss/Track-Anything)]|
-  <br> *Liqi Yan* | <img width="549" alt="image" src="https://user-images.githubusercontent.com/18592211/232305466-ad68546f-b5b1-4c2a-a543-78dea66c7151.png"> | [[Github](https://github.com/ylqi/Count-Anything)]|
-  <br> *Zongxin Yang* | <img width="954" alt="image" src="https://user-images.githubusercontent.com/18592211/232711476-895699e5-fc11-4624-a9fa-e34d84438342.png"> | [[Github](https://github.com/z-x-yang/Segment-and-Track-Anything)]|
-  <br> *Lumin Xu\*, Sheng Jin\*, Wang Zeng, Wentao Liu, Chen Qian, Wanli Ouyang, Ping Luo, Xiaogang Wang* <br> > CUHK, SenseTime <br> > ECCV'22 Oral <br><br> [[**Pose-for-Everything (Project)**](https://github.com/luminxu/Pose-for-Everything)] |  | [[Github](https://github.com/luminxu/Pose-for-Everything)]|
-  <br> Jun Cen, Yizheng Wu, Xingyi Li, Jingkang Yang, Yixuan Pei, Lingdong Kong <br> > Visual Intelligence Lab@HKUST, <br> > HUST, <br> > MMLab@NTU, <br> > Smiles Lab@XJTU, <br> > NUS |  | [Github](https://github.com/Jun-CEN/SegmentAnyRGBD) |
-  <br> *Lumin Xu\*, Sheng Jin\*, Wang Zeng, Wentao Liu, Chen Qian, Wanli Ouyang, Ping Luo, Xiaogang Wang* <br> > CUHK, SenseTime <br> > ECCV'22 Oral <br><br> [[**Pose-for-Everything (Project)**](https://github.com/luminxu/Pose-for-Everything)] |  | [[Github](https://github.com/luminxu/Pose-for-Everything)]|
-  <br> *Jinyu Yang, Mingqi Gao, Zhe Li, Shang Gao, Fangjing Wang, Feng Zheng* <br><br> [[Track-Anything (Project)](https://github.com/gaomingqi/Track-Anything)] |  | [[Github](https://github.com/gaomingqi/Track-Anything)] <br> [[Demo](https://huggingface.co/spaces/watchtowerss/Track-Anything)]|
-  <br> Jun Cen, Yizheng Wu, Xingyi Li, Jingkang Yang, Yixuan Pei, Lingdong Kong <br> > Visual Intelligence Lab@HKUST, <br> > HUST, <br> > MMLab@NTU, <br> > Smiles Lab@XJTU, <br> > NUS |  | [Github](https://github.com/Jun-CEN/SegmentAnyRGBD) |
-  <br> *Achal Dave, Pavel Tokmakov, Deva Ramanan* <br> > ICCV'19 Workshop <br><br> [[**segment-any-moving (Project)**](https://github.com/achalddave/segment-any-moving)] | [<img src="http://www.achaldave.com/projects/anything-that-moves/videos/ZXN6A-tracked-with-objectness-trimmed.gif" width="32%" />](http://www.achaldave.com/projects/anything-that-moves/videos/ZXN6A-tracked-with-objectness-trimmed.mp4)[<img src="http://www.achaldave.com/projects/anything-that-moves/videos/c95cd17749.gif" width="32%" />](http://www.achaldave.com/projects/anything-that-moves/videos/c95cd17749.mp4)<img src="http://www.achaldave.com/projects/anything-that-moves/videos/e0bdb5dfae.gif" width="32%" /> | [[Github](https://github.com/achalddave/segment-any-moving)] |
-
AnyGeneration
-  <br> *Robin Rombach and Andreas Blattmann and Dominik Lorenz and Patrick Esser and Björn Ommer* <br> > LMU München, Runway ML <br> > CVPR'22 <br><br> [[**Stable-Diffusion (Project)**](https://github.com/CompVis/stable-diffusion)] |  | [[Github](https://github.com/CompVis/stable-diffusion)] <br> [[Page](https://stablediffusionweb.com/)] <br> [[Demo](https://stablediffusionweb.com/#demo)] |
-  <br> *Lvmin Zhang, Maneesh Agrawala* <br> > Stanford University <br> > Preprint'23 <br><br> [[**ControlNet (Project)**](https://github.com/lllyasviel/ControlNet)] |  | [[Github](https://github.com/lllyasviel/ControlNet)] <br> [[Demo](https://huggingface.co/spaces/hysts/ControlNet)] |
-  <br> *Tao Yu* |  | [[Github](https://github.com/geekyutao/Inpaint-Anything)] |
- ] |
-  <br> *Shanghua Gao, Pan Zhou* |  | [[Github](https://github.com/sail-sg/EditAnything)] |
-  <br> *Chengsong Zhang* | <img width="659" alt="image" src="https://user-images.githubusercontent.com/18592211/231410895-eac4c4b6-ee61-487b-9333-8dcd1befc610.png"> | [[Github](https://github.com/continue-revolution/sd-webui-segment-anything)] |
-  <br> *Jinwoo Park* |  | [[Github](https://github.com/Curt-Park/segment-anything-with-clip)] |
-  <br> *LV-Lab, NUS* |  | [Github](https://github.com/Huage001/Transfer-Any-Style) |
-  <br> *Zeqiang-Lai* |  | [Github](https://github.com/Zeqiang-Lai/Anything2Image) |
-  <br> *Zhihang Zhong, Gurunandan Krishnan, Xiao Sun, Yu Qiao, Sizhuo Ma, Jian Wang* <br> > Shanghai AI Laboratory, Snap Inc. <br> > Preprint'23 |  | [[Github]](https://github.com/zzh-tech/InterpAny-Clearer) <br> [[Page]](https://zzh-tech.github.io/InterpAny-Clearer/) <br> [[ArXiv]](https://arxiv.org/abs/2311.08007) |
- High-Resolution Image Synthesis with Latent Diffusion Models - to-Image Generation |
- Adding Conditional Control to Text-to-Image Diffusion Models
- Inpaint Anything: Segment Anything Meets Image Inpainting
- ] |
-  <br> *Shanghua Gao, Pan Zhou* |  | [[Github](https://github.com/sail-sg/EditAnything)] |
-  <br> *Robin Rombach and Andreas Blattmann and Dominik Lorenz and Patrick Esser and Björn Ommer* <br> > LMU München, Runway ML <br> > CVPR'22 <br><br> [[**Stable-Diffusion (Project)**](https://github.com/CompVis/stable-diffusion)] |  | [[Github](https://github.com/CompVis/stable-diffusion)] <br> [[Page](https://stablediffusionweb.com/)] <br> [[Demo](https://stablediffusionweb.com/#demo)] |
-  <br> *Chengsong Zhang* | <img width="659" alt="image" src="https://user-images.githubusercontent.com/18592211/231410895-eac4c4b6-ee61-487b-9333-8dcd1befc610.png"> | [[Github](https://github.com/continue-revolution/sd-webui-segment-anything)] |
-  <br> *Jinwoo Park* |  | [[Github](https://github.com/Curt-Park/segment-anything-with-clip)] |
- GigaGAN: Large-scale GAN for Text-to-Image Synthesis - scale GAN |
-
Any3D
-  <br> *Zhening Huang, Xiaoyang Wu, Xi Chen, Hengshuang Zhao, Lei Zhu, Joan Lasenby* <br> > Cambridge, HKU, HKUST <br><br> [[**OpenIns3D**](https://github.com/Pointcept/OpenIns3D)] |  | [[Github](https://github.com/Pointcept/OpenIns3D)] <br> [[Page](https://zheninghuang.github.io/OpenIns3D/)]|
-  <br> *LV-Lab, NUS* |  <br>  | [Github](https://github.com/Anything-of-anything/Anything-3D) |
-  <br> *Nexuslrf* |  | [Github](https://github.com/nexuslrf/SAM-3D-Selector) |
-  <br> *dvlab-research* |  | [[Github](https://github.com/dvlab-research/3D-Box-Segment-Anything)] |
-  <br> *Gongfan Fang, Xinyin Ma, Mingli Song, Michael Bi Mi, Xinchao Wang* <br> > Learning and Vision Lab @ NUS<br> > CVPR'23 <br><br> [[**Torch-Pruning (Project)**](https://github.com/VainF/Torch-Pruning)] |  | [[Github](https://github.com/VainF/Torch-Pruning)] <br> [[Demo](https://colab.research.google.com/drive/1TRvELQDNj9PwM-EERWbF3IQOyxZeDepp?usp=sharing)] |
-  <br> *Tianyi Chen, Luming Liang, Tianyu Ding, Ilya Zharkov* <br> > Microsoft <br> > ICLR'23 <br><br> [[**Only Train Once (Project)**](https://github.com/tianyic/only_train_once)] |  | [[Github](https://github.com/tianyic/only_train_once)] |
-  <br> *Xueyan Zou, Zi-Yi Dou, Jianwei Yang, Zhe Gan, Linjie Li, Chunyuan Li, Xiyang Dai, Harkirat Behl, Jianfeng Wang, Lu Yuan, Nanyun Peng, Lijuan Wang, Yong Jae Lee, Jianfeng Gao* <br> > Microsoft <br> > CVPR'23 <br><br> [[**X-Decoder (Project)**](https://github.com/microsoft/X-Decoder/)] |  | [[Github](https://github.com/microsoft/X-Decoder/)] <br> [[Page](https://x-decoder-vl.github.io)] <br> [[Demo](https://huggingface.co/spaces/xdecoder/Demo)] |
- ![Star - noah/Pretrained-IPT) <br> [**Pre-Trained Image Processing Transformer**]() <br> *Chen, Hanting and Wang, Yunhe and Guo, Tianyu and Xu, Chang and Deng, Yiping and Liu, Zhenhua and Ma, Siwei and Xu, Chunjing and Xu, Chao and Gao, Wen* <br> > Huawei-Noah <br> > CVPR'21 <br><br> [[**Pretrained-IPT (Project)**](https://github.com/huawei-noah/Pretrained-IPT)] |  | [[Github](https://github.com/huawei-noah/Pretrained-IPT)] |
- ![Star
- HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in HuggingFace
- Generalized Decoding for Pixel, Image and Language
- Pre-Trained Image Processing Transformer - level Vision |
- TaskMatrix.AI: Completing Tasks by Connecting Foundation Models with Millions of APIs
-
AnyX
Categories
Sub Categories