Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/woheller69/llama_tk_chat
Simple chat interface for local AI using llama-cpp-python and llama-cpp-agent
https://github.com/woheller69/llama_tk_chat
gui llama-cpp-agent llama-cpp-python llm-inference
Last synced: 6 days ago
JSON representation
Simple chat interface for local AI using llama-cpp-python and llama-cpp-agent
- Host: GitHub
- URL: https://github.com/woheller69/llama_tk_chat
- Owner: woheller69
- License: mit
- Created: 2024-05-08T13:13:18.000Z (6 months ago)
- Default Branch: main
- Last Pushed: 2024-07-27T10:44:10.000Z (4 months ago)
- Last Synced: 2024-10-31T08:37:26.091Z (13 days ago)
- Topics: gui, llama-cpp-agent, llama-cpp-python, llm-inference
- Language: Python
- Homepage:
- Size: 127 KB
- Stars: 8
- Watchers: 4
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# LLAMA_TK_CHAT
Simple chat interface for local AI using llama-cpp-python and llama-cpp-agent## Quickstart
The TK GUI is based on `llama-cpp-python`, `llama-cpp-agent`, `typer`, and `tkinter` package
(install them with pip install ...).The GUI is a self-contained Python script named `LLAMA_TK_GUI.py`. As long as
its package dependencies are present, you can download and run it from wherever you like.Specify the path to the model with the `-m`/`--model` argument and the prompt template with `-f`/`--format` (default: `CHATML`, Options: MISTRAL, CHATML, VICUNA, LLAMA_2, SYNTHIA, NEURAL_CHAT, SOLAR, OPEN_CHAT, ALPACA, CODE_DS, B22, LLAMA_3, PHI_3, OPEN_INTERPRETER, AUTOCODER, GEMMA_2, DEEP_SEEK_CODER_2)
A optional system prompt can be set with `-s`/`--sysprompt`.
Context length defaults to 2048, it can be set with `-c`/`--context-length`. The number of threads can be set with `-t`/`--n-threads` (default: 4 threads). Temperature can be set with `--temperature`, default is 0.65.
`-l`/`--mlock` allows to use MLOCK instead of MMAP.
Everything runs on CPU.Input your request in the bottom window and click ```Generate```.
Inference can be interrupted using the ``Stop`` button.