Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/cheshire-cat-ai/local-cat
A ready to use setup for Cat + Ollama + Embedder + Qdrant
https://github.com/cheshire-cat-ai/local-cat
Last synced: about 1 month ago
JSON representation
A ready to use setup for Cat + Ollama + Embedder + Qdrant
- Host: GitHub
- URL: https://github.com/cheshire-cat-ai/local-cat
- Owner: cheshire-cat-ai
- License: gpl-3.0
- Created: 2023-11-08T17:09:20.000Z (about 1 year ago)
- Default Branch: main
- Last Pushed: 2024-06-14T11:39:16.000Z (6 months ago)
- Last Synced: 2024-08-17T08:07:09.072Z (4 months ago)
- Size: 99.6 KB
- Stars: 42
- Watchers: 2
- Forks: 15
- Open Issues: 2
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# local-cat πΈπ
**local-cat** provides a completely local setup for CheshireCat. Local-cat leverages Local runners + Qdrant to run your preferred LLM, Embedder and VectorDB locally.
> [!WARNING]
>
> - **Technical Expertise Required:** Setting up and running local-cat requires some technical know-how.
> - **Hardware Requirements:** Performance may be slow without a recent GPU or NPU.## Ollama Setup
> [!IMPORTANT]
> Ollama can be instable with **latest models** or **non-common use** models(like qwen, deepseek)!!
> If you encount inference problems, downgrade ollama image or [open an issue to Ollama](https://github.com/ollama/ollama/issues)### Setup Instructions
1. **Clone the Repository:** `git clone https://github.com/cheshire-cat-ai/local-cat.git`
2. **Navigate to the Directory:** `cd local-cat`
3. **Start local-cat:** `docker-compose up -d`
4. **Pull Your Desired Model:** `docker exec ollama_cat ollama pull `
- Replace `` with the specific model you want to use.
5. **Your Setup is Complete!**
- You can now install additional plugins or start interacting with local-cat.### Use Ollama with MacOS GPU Acceleration
Ollama normally handles running the model with GPU acceleration. In order to use GPU acceleration on Mac OS it is recommended to run Ollama directly on the host machine rather than inside Docker. More info [here](https://ollama.com/blog/ollama-is-now-available-as-an-official-docker-image).
> [!NOTE]
> This is recommended until GPU acceleration is supported by Docker Desktop on MacOS.To use local-cat with GPU acceleration on Mac:
1. Install the menu bar app version of Ollama, which is the current recommended setup for MacOS users.
2. Start using the following command `docker compose -f docker-compose-macos.yml up`
3. Configure the Ollama Base URL in the cat's LLM settings to `http://host.docker.internal:11434`.> Note: This configuration allows Docker containers to communicate with your locally running Ollama service and leverage MacOS GPU acceleration.
### Use Ollama with AMD
To use local-cat with [AMD graphics that supports ROCm](https://rocm.docs.amd.com/en/docs-5.7.0/release/gpu_os_support.html#linux-supported-gpus), use the following command:
```bash
docker compose -f docker-compose-amd.yml up
```