An open API service indexing awesome lists of open source software.

https://github.com/ezforever/llama.cpp-static

Static builds of llama.cpp (Currently only amd64 server builds are available)
https://github.com/ezforever/llama.cpp-static

llama llama-cpp llamacpp localai self-hosted

Last synced: 3 months ago
JSON representation

Static builds of llama.cpp (Currently only amd64 server builds are available)

Awesome Lists containing this project

README

        

# llama.cpp-static
Your daily, minimal build of llama.cpp. [Also available on Docker Hub.](https://hub.docker.com/r/ezforever/llama.cpp-static)

Source code:
Built from:

## Usage

Please refer to [llama.cpp docker guide](https://github.com/ggerganov/llama.cpp/blob/master/README.md#docker) and [server README](https://github.com/ggerganov/llama.cpp/tree/master/examples/server).

## Tag format

**tl;dr:** Use `server-ssl-avx2` if you don't know what you're doing.

Server images are tagged in the format of `server--`.

`` is one of the following:

- `nossl`: Minimal build with no SSL/TLS capability.
- `ssl`: Built with OpenSSL (`LLAMA_SERVER_SSL=ON`), thus supports `--ssl-key-file` and `--ssl-cert-file`.

`` is one of the following:

- `noavx`: All AVX-related optimizations are disabled. Do not use this build unless you are working around some known bug, or running LLMs on a 10-year-old potato.
- `avx`: (Only) AVX instruction set is enabled. Might be useful if you are using some old CPUs that don't support AVX2.
- `avx2`: AVX2 instruction set is enabled. This build should support most modern/recent CPUs with reasonable performance.
- `avx512`: AVX512 base instruction set is enabled. Currently only some high-end or server-grade CPUs support this instruction set, so check your hardware specs before using this build.
- `oneapi`: Experimental build with the Intel oneAPI compiler, inspired by ggerganov/llama.cpp#5067. Offers a ~30% speed increase (~20tok/s vs ~15tok/s) in prompt processing on my machine compared to `avx2` builds. Not updated daily.

RPC server images are tagged in the format of `rpc-server--`. Refer to [rpc README](https://github.com/ggerganov/llama.cpp/tree/master/examples/rpc) for detailed information.