https://github.com/thoughtscript/hugging_face_llm_2025
Hugging Face LLM on Fast API exploration
https://github.com/thoughtscript/hugging_face_llm_2025
docker fastapi llm machine-learning python tiny-llm
Last synced: 8 months ago
JSON representation
Hugging Face LLM on Fast API exploration
- Host: GitHub
- URL: https://github.com/thoughtscript/hugging_face_llm_2025
- Owner: Thoughtscript
- Created: 2025-05-11T02:47:19.000Z (9 months ago)
- Default Branch: main
- Last Pushed: 2025-05-21T02:31:41.000Z (8 months ago)
- Last Synced: 2025-05-21T03:33:16.166Z (8 months ago)
- Topics: docker, fastapi, llm, machine-learning, python, tiny-llm
- Language: HTML
- Homepage:
- Size: 13.7 KB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Hugging_Face_LLM_2025
[](https://www.python.org/downloads/)
[](https://www.docker.com/)
[](https://huggingface.co/arnir0/Tiny-LLM)
> Experimenting with: **Tiny Large Language Models**.
## Setup and Use
```bash
docker compose up
```
1. http://localhost:8000/public/index.html
2. http://localhost:8000/api/llm?prompt=abcdefghijklmnopqrstuvwxyz
## Warning
1. This is a simple example (**NOT** Production-worthy) and basic safeguards (like input field validation, param sanitization, and the like) are mostly omitted here.
* There's not a tremendous one can do within the context of *this* simple demo to break things/be malicious.
* But the design pattern *should* be avoided in Production *as is* (without the addition of typical Production security mechanisms)!
1. The build can take upwards of `30 minutes` due to large (`many GB`) LLM Models.
## Notes
1. Hugging Face will download and cache `models--arnir0--Tiny-LLM` into `root/.cache/huggingface/hub`.
2. Also, **Docker Desktop 4.40** now supports [Docker Model Runner](https://www.docker.com/blog/introducing-docker-model-runner/).
* The following commands will launch of Dockerized LLM Model ([somewhat similar](https://hub.docker.com/r/ai/deepseek-r1-distill-llama) to this one):
* `docker model pull ai/deepseek-r1-distill-llama`
* `docker model run ai/deepseek-r1-distill-llama`
* **Docker Model Runner** supports Hugging Face Models.
## Resources and Links
1. https://huggingface.co/arnir0/Tiny-LLM
2. https://huggingface.co/learn/llm-course/