Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/robgon-art/open-clip

Test out OpenCLIP for Image Search and Automatic Captioning
https://github.com/robgon-art/open-clip

Last synced: about 2 months ago
JSON representation

Test out OpenCLIP for Image Search and Automatic Captioning

Awesome Lists containing this project

README

        

# **Using OpenCLIP for Image Search and Automatic Captioning**
## How LAION used more data and new ML training techniques to improve image and text embeddings for various applications

![open-clip cover Image](https://raw.githubusercontent.com/robgon-art/open-clip/main/cover_med.jpg)

**By Robert. A Gonsalves**

You can see my article on Medium.

The source code and generated images are released under the [CC BY-SA license](https://creativecommons.org/licenses/by-sa/4.0/).
![CC BYC-SA](https://licensebuttons.net/l/by-sa/3.0/88x31.png)

## Google Colabs
* [Test CLIP with Library of Congress Images](https://colab.research.google.com/github/robgon-art/open-clip/blob/main/Test_LOC_Photos_with_CLIP.ipynb)
* [Test OpenCLIP with Library of Congress Images](https://colab.research.google.com/github/robgon-art/open-clip/blob/main/Test_LOC_Photos_with_OpenCLIP.ipynb)
* [Index Library of Congress Photos with OpenCLIP](https://colab.research.google.com/github/robgon-art/open-clip/blob/main/Index_LOC_Photos_with_OpenCLIP.ipynb)
* [Search for Library of Congress Photos with OpenCLIP](https://colab.research.google.com/github/robgon-art/open-clip/blob/main/Search_for_LOC_Photos_with_OpenCLIP.ipynb)
* [Create Captions with CoCa and OpenCLIP](https://colab.research.google.com/github/robgon-art/open-clip/blob/main/Create_Captions_with_CoCa_and_OpenCLIP.ipynb)

## Acknowledgements
---
- A. Radford et al., CLIP, Learning Transferable Visual Models From Natural Language Supervision (2021)
- M. Cherti et al., OpenCLIP, Reproducible scaling laws for contrastive language-image learning (2022)
- G. Couairon et al., Embedding Arithmetic of Multimodal Queries for Image Retrieval (2022)
- S. Wang and P. Kanwar, BFloat16: The secret to high performance on Cloud TPUs (2019)
- J. Yu, CoCa: Contrastive Captioners are Image-Text Foundation Models (2022)

---

To get unlimit

## Citation
To cite this repository:

```bibtex
@software{GreenLIT,
author = {Gonsalves, Robert A.},
title = {Using OpenCLIP for Image Search and Automatic Captioning},
url = {https://github.com/robgon-art/open-clip},
year = 2023,
month = February
}
```