Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
awesome-self-supervised-multimodal-learning
A curated list of self-supervised multimodal learning resources.
https://github.com/ys-zong/awesome-self-supervised-multimodal-learning
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - research/tree/master/mmv)
- [paper - NCE_HowTo100M)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - research/big_vision)
- [paper
- [paper
- [paper
- [paper - science/crossmodal-contrastive-learning)
- [paper
- [paper
- [paper - research/google-research/tree/master/vatt)
- [paper
- [paper - morgado/AVSpatialAlignment)
- [paper
- [paper
- [paper - Part-of-Speech-Embeddings)
- [paper
- [paper
- [paper
- [paper - CMA)
- [paper
- [paper
- [paper
- [paper - of-Pixels)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - label)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - io-inference)
- [paper - beit)
- [paper - ai/models/tree/master/research/mm/opt)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - multi-task)
- [paper
- [paper
- [paper - bug/mpre-unmasked)
- [paper
- [paper
- [paper - Clustering-Network)
- [paper
- [paper - videotext)
- [paper
- [paper
- [paper
- [paper - action-conditional-video-prediction)
- [paper
- [paper - research/planet)
- [paper
- [paper - forward-model)
- [paper
- [paper
- [paper
- [paper - group/CoMIR)
- [paper - pytorch)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - supervised_change_detetction)
- [paper
- [paper
- [paper
- [paper - machine-translation-using-monolingual-corpora-only-pytorch)
- [paper
- [paper - grounding)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - distill.cs.uni-freiburg.de/)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper - Neural-Scaling)
- [paper - GVT/DeCLIP)
- [paper - language-models-are-bows)
- [paper
- [paper - ood)
- [paper - jia/BadEncoder)
- [paper - science/multimodal-robustness)
- [paper
- [paper
- [paper - Liang/Modality-Gap)
- [paper
- [paper
- [paper
- [paper
- [paper
- [paper
- Link - caption)|
- Link - |
- Link
- Link
- Link - xirong/flickr8kcn)|
- Link - |
- Link
- Link - research-datasets/conceptual-captions)|
- Link
- Github
- Link - |
- Link - |
- Link - |
- Link - Med-2019)|
- Link
- Link - vcr)|
- Details - ml/SNLI-VE)|[Github](https://github.com/necla-ml/SNLI-VE)|
- Link - lab/nlvr)|
- Link - lab/nlvr)|
- Link - |
- Link
- Link
- Link
- Link
- Link - narratives)|
- Link
- Link - video-captioning-pytorch)|
- Link
- Link - xw/Video-guided-Machine-Translation)|
- Link - YouCook2)|
- Link
- Link - vtt-it)|
- Link - |
- Link - Release)|
- Link - |
- Link
- Link
- Link - dataset/annotations)|
- Link - |
- Link - |
- Link
- Link
- Link - of-Pixels)|
- Link - to-Listen-at-the-Cocktail-Party)|
- Link - |
- Link
- Link - ECCV18)|
- Link - morgado/spatialaudiogen)|
- Link - Perceptual-Computing-Lab/openpose)|
- Link
- Link - |
- Link
- Link - |
- Link - vgd/scanobjectnn)|
- Link - |
- Link - devkit)|
- Link - kitti-api)|