Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/karaketir16/openwebui-langfuse
This repository provides resources and guidelines to facilitate the integration of Open-WebUI and Langfuse, enabling seamless monitoring and management of AI model usage statistics.
https://github.com/karaketir16/openwebui-langfuse
docker docker-compose langfuse llm llm-webui ollama ollama-webui open-webui pipelines
Last synced: 27 days ago
JSON representation
This repository provides resources and guidelines to facilitate the integration of Open-WebUI and Langfuse, enabling seamless monitoring and management of AI model usage statistics.
- Host: GitHub
- URL: https://github.com/karaketir16/openwebui-langfuse
- Owner: karaketir16
- License: mit
- Created: 2024-06-23T11:18:29.000Z (5 months ago)
- Default Branch: master
- Last Pushed: 2024-06-23T11:39:41.000Z (5 months ago)
- Last Synced: 2024-06-23T12:33:56.851Z (5 months ago)
- Topics: docker, docker-compose, langfuse, llm, llm-webui, ollama, ollama-webui, open-webui, pipelines
- Language: Python
- Homepage:
- Size: 6.84 KB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.MD
- License: LICENSE
Awesome Lists containing this project
README
# OLLAMA + OPEN-WEBUI + PIPELINES + LANGFUSE
## Introduction
This repository provides a setup for integrating [OLLAMA](https://github.com/ollama/ollama), [OPEN-WEBUI](https://github.com/open-webui/open-webui), [PIPELINES](https://github.com/open-webui/pipelines/), and [LANGFUSE](https://github.com/langfuse/langfuse) using Docker. Follow the steps below to get everything up and running.
## Prerequisites
- Docker and required GPU drivers installed on your system.
## Installation
1. **Clone this repository:**
```bash
git clone https://github.com/karaketir16/openwebui-langfuse.git
cd openwebui-langfuse
```2. **Run the setup script:**
```bash
./run-compose.sh
```
or
```bash
docker compose up -d
# default driver is nvidia
```## Configuration
### Langfuse Setup
1. **Download the `langfuse_filter_pipeline.py` file:**
- You can download it from [here](https://github.com/open-webui/pipelines/blob/main/examples/filters/langfuse_filter_pipeline.py), or use the provided file in `example/langfuse_filter_pipeline.py`.2. **Access Langfuse:**
- Open your browser and go to `http://localhost:4000`.3. **Create an Admin Account and Project:**
- Create an admin account and then create a project.
- Go to Project Settings and create an API key.
- Retrieve the secret key and public key.4. **Update the Pipeline Script:**
- Replace `your-secret-key-here` with your secret key, `your-public-key-here` with your public key, and `https://cloud.langfuse.com` with `http://langfuse:4000` in `langfuse_filter_pipeline.py`.
- **Note:** If you are using the example in this repository, replace the secret key and public key with your own keys.### Open-WebUI Setup
1. **Access Open-WebUI:**
- Open your browser and go to `http://localhost:3000`.2. **Create an Admin Account:**
- Create an admin account.3. **Upload the Pipeline Script:**
- Navigate to `Settings -> Admin Settings -> Pipelines`.
- In the `Upload Pipeline` section, select the `langfuse_filter_pipeline.py` file and click the upload button.4. **Monitor Usage:**
- You can now monitor Open-WebUI usage statistics from Langfuse.### Model Downloading
1. **Access Open-WebUI:**
- Open your browser and go to `http://localhost:3000`.2. **Create an Admin Account:**
- Create an admin account.3. **Pull Models:**
- Navigate to `Settings -> Admin Settings -> Models`.
- Enter a model tag to pull from the Ollama library (e.g., `phi3:mini`).
- Press the pull button.