https://github.com/the-crypt-keeper/llamacpp-horde-bridge
Use llama.cpp server to join the AI Horde
https://github.com/the-crypt-keeper/llamacpp-horde-bridge
Last synced: 2 months ago
JSON representation
Use llama.cpp server to join the AI Horde
- Host: GitHub
- URL: https://github.com/the-crypt-keeper/llamacpp-horde-bridge
- Owner: the-crypt-keeper
- License: agpl-3.0
- Created: 2023-12-21T21:49:33.000Z (over 1 year ago)
- Default Branch: master
- Last Pushed: 2024-02-09T14:01:59.000Z (over 1 year ago)
- Last Synced: 2025-01-28T20:46:53.942Z (4 months ago)
- Language: Python
- Size: 4.53 MB
- Stars: 1
- Watchers: 3
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- Funding: .github/FUNDING.yml
- License: LICENSE
Awesome Lists containing this project
README
# LlamaCpp Horde Bridge
This software enables you to join your [llama.cpp](https://github.com/ggerganov/llama.cpp) server to the [KoboldAI Horde](https://github.com/db0/AI-Horde) and make it into a Scribe worker, performing distributed text generation.
It is a fork of [KoboldAI-Horde-Bridge](https://raw.githubusercontent.com/db0/KoboldAI-Horde-Bridge).
# Why llama.cpp and not koboldcpp?
See [this reddit post](https://www.reddit.com/r/LocalLLaMA/comments/18helbs/how_to_run_mixtral_8x7b_gguf_on_tesla_p40_without/), using this trick older Pascal GPUs (GTX 10x0, P40, K80) are almost twice as fast, particulary at long contexts.
Compile [llama.cpp](https://github.com/ggerganov/llama.cpp) with `make LLAMA_CUBLAS=1 LLAMA_CUDA_FORCE_MMQ=1` to get a Pascal-optimized `server` binary.
# Instructions
- Launch llama.cpp server, something like: `server -m /path/to/model.gguf -ngl 100 -c 2048`
- Obtain a [Horde API key](https://aihorde.net/register)
- Copy `clientData_template.py` to `clientData.py` and customize the configuration:
- `kai_url` LlamaCpp server endpoint (default OK if same machine)
- `kai_name` Horde worker name
- `api_key` Hode API key
- Run `bridge.py`Note that for quick Testing, you can provide these arguments via the CLI: `bridge.py -k -a -n `