Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/kleinyuan/llama2-csv-webapp
self host/local host llama2 based web app to chat with your csvs (multiple)
https://github.com/kleinyuan/llama2-csv-webapp
chatgpt csv datalake large-language-models llama2 llm meta openai pandas pandas-ai pandasai streamlit
Last synced: about 2 months ago
JSON representation
self host/local host llama2 based web app to chat with your csvs (multiple)
- Host: GitHub
- URL: https://github.com/kleinyuan/llama2-csv-webapp
- Owner: KleinYuan
- License: mit
- Created: 2024-01-15T02:00:24.000Z (12 months ago)
- Default Branch: main
- Last Pushed: 2024-01-15T02:07:26.000Z (12 months ago)
- Last Synced: 2024-10-18T22:04:45.939Z (3 months ago)
- Topics: chatgpt, csv, datalake, large-language-models, llama2, llm, meta, openai, pandas, pandas-ai, pandasai, streamlit
- Language: Python
- Homepage:
- Size: 168 KB
- Stars: 6
- Watchers: 3
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# llama2-csv-webapp
Show case running a couple of lines of code to
- launch a web app
- with local hosted LLM, such as llama2, without exposing your data to external commercial products.
- chat about csvI am trying to keep the code as simple / naive as possible to make it easy enough to read, instead of optimizing the robustness.
## Machine Requirements
The >= 24GB Memory GPUs (such as 3090ti, 4090, ...) is all you need. If your GPU memory is lower than that, just swap the local model with a smaller one.
Everything below is tested on a Ubuntu 22.04 machine, in the native system with Anaconda.
## Step by Step Guide
### Install Python Dedpendencies
`pip install -r requirements.txt`
### Spin Up Local LLM Inference Server
Follow [this](https://huggingface.co/docs/text-generation-inference/quicktour#quick-tour) or run the follows:
```
model=codellama/CodeLlama-7b-Instruct-hf
volume=$PWD/data
docker run --gpus all --shm-size 1g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:1.3 --model-id $model
```(You can change `codellama/CodeLlama-7b-Instruct-hf` to smaller models if your GPU has low Memory)
### Spin Up Streamlit Web App
`streamlit run app.py`
Open `http://10.0.0.171:8501/` on your browser and you are all set!
![demo](demo.png)