Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/ivangabriele/docker-llm
Pre-loaded LLMs served as an OpenAI-Compatible API via Docker images.
https://github.com/ivangabriele/docker-llm
api docker docker-image llm llms llong lmsys openai openai-api openorca orca runpod server vast vicuna vllm
Last synced: about 2 months ago
JSON representation
Pre-loaded LLMs served as an OpenAI-Compatible API via Docker images.
- Host: GitHub
- URL: https://github.com/ivangabriele/docker-llm
- Owner: ivangabriele
- License: agpl-3.0
- Created: 2023-10-15T17:07:33.000Z (about 1 year ago)
- Default Branch: main
- Last Pushed: 2023-10-31T07:51:44.000Z (about 1 year ago)
- Last Synced: 2024-10-12T19:32:44.699Z (2 months ago)
- Topics: api, docker, docker-image, llm, llms, llong, lmsys, openai, openai-api, openorca, orca, runpod, server, vast, vicuna, vllm
- Language: Dockerfile
- Homepage:
- Size: 199 KB
- Stars: 5
- Watchers: 1
- Forks: 2
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- Contributing: CONTRIBUTING.md
- License: LICENSE.md
- Code of conduct: CODE_OF_CONDUCT.md
- Security: SECURITY.md
Awesome Lists containing this project
README
# Docker Image ―
OpenAI API-Compatible Pre-loaded LLM Server[![img-github]][link-github]
[![img-docker]][link-docker]Docker images are based on [Nvidia CUDA images](https://hub.docker.com/r/nvidia/cuda).
LLMs are pre-loaded and served via [vLLM](https://github.com/vllm-project/vllm).## Environment Variables
- `TENSOR_PARALLEL_SIZE`: Number of GPUs to use. Default: `1`.
## Port
The OpenAI API is exposed on port `8000`.
## Tags & Deployment Links
> [!NOTE]
> The VRAM column is the minimum required amount of VRAM used by the model on a single GPU.| Tag | Model | RunPod | Vast.ai | VRAM |
| ----------------------------------------------- | ------------------------------------------------------------------------ | ----------------------------------------------------------------------- | ------------------------------------------------------------------------------------------------- | ---- |
| `ivangabriele/llm:lmsys__vicuna-13b-v1.5-16k` | [![img-huggingface]](https://huggingface.co/lmsys/vicuna-13b-v1.5-16k) | [![img-runpod]](https://runpod.io/gsc?template=7qj5x0er29&ref=s0k66ov1) | [![img-vastai]](https://cloud.vast.ai/?ref_id=46067&template_id=2e645ca0121df6858ca25e73938d118e) | 26GB |
| `ivangabriele/llm:open-orca__llongorca-13b-16k` | [![img-huggingface]](https://huggingface.co/Open-Orca/LlongOrca-13B-16k) | [![img-runpod]](https://runpod.io/gsc?template=b4fx781ink&ref=s0k66ov1) | [![img-vastai]](https://cloud.vast.ai/?ref_id=46067&template_id=6ab428f190858336446c2ee38ceb2cc6) | 26GB |## Roadmap
- [ ] Add more popular models.
- [ ] Start the server in background to allow for SSH access.---
[img-docker]: https://img.shields.io/docker/pulls/ivangabriele/llm?style=for-the-badge
[img-github]: https://img.shields.io/badge/Github-Repo-black?logo=github&style=for-the-badge
[img-github-actions]: https://img.shields.io/github/actions/workflow/status/ivangabriele/docker-llm/main.yml?branch=main&style=for-the-badge
[img-huggingface]: https://img.shields.io/badge/README-ffd21e?style=for-the-badge
[img-runpod]: https://img.shields.io/badge/Deploy-673ab7?style=for-the-badge
[img-vastai]: https://img.shields.io/badge/Deploy-000000?style=for-the-badge[link-docker]: https://hub.docker.com/r/ivangabriele/llm
[link-github]: https://github.com/ivangabriele/docker-llm
[link-github-actions]: https://github.com/ivangabriele/docker-llm/actions/workflows/main.yml