Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/wangxiao5791509/MultiModal_BigModels_Survey
[MIR-2023-Survey] A continuously updated paper list for multi-modal pre-trained big models
https://github.com/wangxiao5791509/MultiModal_BigModels_Survey
anhui-university audio big-models depth event-camera multi-modal natural-language pengchenglab point-cloud pre-training radar review rgb-text-audio self-attention survey thermal-infrared transformers
Last synced: 4 months ago
JSON representation
[MIR-2023-Survey] A continuously updated paper list for multi-modal pre-trained big models
- Host: GitHub
- URL: https://github.com/wangxiao5791509/MultiModal_BigModels_Survey
- Owner: wangxiao5791509
- Created: 2021-12-13T09:19:44.000Z (over 2 years ago)
- Default Branch: main
- Last Pushed: 2024-03-16T08:18:16.000Z (4 months ago)
- Last Synced: 2024-03-16T09:31:21.816Z (4 months ago)
- Topics: anhui-university, audio, big-models, depth, event-camera, multi-modal, natural-language, pengchenglab, point-cloud, pre-training, radar, review, rgb-text-audio, self-attention, survey, thermal-infrared, transformers
- Homepage:
- Size: 12.3 MB
- Stars: 244
- Watchers: 8
- Forks: 16
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Lists
- awesome-llm-blogs - MultiModal_BigModels_Survey
README
## This github will be continuously updated for the survey paper:
**Large-scale Multi-Modal Pre-trained Models: A Comprehensive Survey**, [Xiao Wang](https://wangxiao5791509.github.io/), [Guangyao Chen](https://icgy96.github.io/), Guangwu Qian, Pengcheng Gao, [Xiao-Yong Wei](https://scholar.google.com/citations?user=8kxWTokAAAAJ&hl=zh-CN&oi=ao), [Yaowei Wang](https://scholar.google.com/citations?user=o_DllmIAAAAJ&hl=zh-CN&oi=ao), [Yonghong Tian](https://scholar.google.com/citations?user=fn6hJx0AAAAJ&hl=zh-CN&oi=ao), [Wen Gao](https://scholar.google.com/citations?user=b0vWahYAAAAJ&hl=zh-CN&oi=ao).
[[arXiv](https://arxiv.org/abs/2302.10035)]
[[MIR](https://www.mi-research.net/article/doi/10.1007/s11633-022-1410-8)]
[[极市平台公众号](https://mp.weixin.qq.com/s/5eELXfACI67yZT7WUtMFMA)]
[[机器智能研究MIR(MIR编辑部)](https://mp.weixin.qq.com/s/yX1DdDCA-nMluzOB6Qz3sw)]
[[Machine Intelligence Research (Youtube)](https://youtu.be/zQxV-SUz6zU?si=e27cyVjMUdU-XEwd)]------
## Framework of this survey
## Review and Surveys
Please check this file [[Surveys.md](https://github.com/wangxiao5791509/MultiModal_BigModels/blob/main/Surveys.md)]## Datasets
Please check this file [[Datasets.md](https://github.com/wangxiao5791509/MultiModal_BigModels/blob/main/Datasets.md)]## Publications
Please check this file [[paperList.md](https://github.com/wangxiao5791509/MultiModal_BigModels/blob/main/paperList.md)]## Experimental Analysis
## :page_with_curl: BibTex:
If you find this survey useful for your research, please cite the following papers:```bibtex
@article{wang2022MMPTMSurvey,
title={Large-scale Multi-Modal Pre-trained Models: A Comprehensive Survey},
author={Wang, Xiao and Chen, Guangyao and Qian, Guangwu and Gao, Pengcheng and Wei, Xiao-Yong and Wang, Yaowei and Tian, Yonghong and Gao, Wen},
url={https://github.com/wangxiao5791509/MultiModal_BigModels_Survey},
year={2022}
}```
If you have any questions about this survey, please email me via: [email protected] or [email protected]