Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/duydvu/triton-inference-server-web-ui
Triton Inference Server Web UI
https://github.com/duydvu/triton-inference-server-web-ui
ai triton-inference-server user-interface web
Last synced: 3 months ago
JSON representation
Triton Inference Server Web UI
- Host: GitHub
- URL: https://github.com/duydvu/triton-inference-server-web-ui
- Owner: duydvu
- License: mit
- Created: 2023-11-04T16:13:09.000Z (over 1 year ago)
- Default Branch: main
- Last Pushed: 2023-11-06T14:49:47.000Z (over 1 year ago)
- Last Synced: 2024-11-11T23:09:03.657Z (3 months ago)
- Topics: ai, triton-inference-server, user-interface, web
- Language: TypeScript
- Homepage:
- Size: 210 KB
- Stars: 10
- Watchers: 1
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Triton Inference Server Web UI
Manage and monitor your Triton Inference Server with a web browser. You can use this project to develop and test your models on Triton Inference Server more easily. It is also useful for monitoring your models in production but it is not recommended to edit model configuration in production since it will be lost when the server restarts.
This project is built on top of Triton Inference Server HTTP API. It is not affiliated with NVIDIA.
![](resources/screenshot_repo.jpeg)
![](resources/screenshot_model.jpeg)## Features
- List all models from the model repository.
- Easily load/unload models from the model repository.
- View & edit model configuration on the fly.
- Monitor models metrics.## Installation
### Docker
Pull the image:
```bash
docker pull duyvd/triton-inference-server-web-ui
```Run the container:
```bash
docker run -d --name triton-inference-server-web-ui \
-p 3000:3000 \
-e API_URL= \
-e API_AUTH_HEADER= \
duyvd/triton-inference-server-web-ui
```Where:
- `` is the URL of the HTTP server, usually on port 8000.
- `` is optional, if your Triton Inference Server is protected by an authentication header.For local Triton Inference Server, you can use the following command:
```bash
docker run -it --name triton-inference-server-web-ui \
-p 3000:3000 \
-e API_URL=http://localhost:8000 \
duyvd/triton-inference-server-web-ui
```Finally, open your browser and go to http://localhost:3000.
### Build from source
Prerequisites:
- Node.js >= 18Clone the repository and install dependencies:
```bash
git clone https://github.com/duydvu/triton-inference-server-web-ui
cd triton-inference-server-web-ui
npm install
```Prepare the `.env` file at the root of the project:
```bash
# The URL of the HTTP server, usually on port 8000
API_URL=
# Optional, if your Triton Inference Server is protected by an authentication header
API_AUTH_HEADER=
```For local Triton Inference Server, you can use the following `.env` file:
```bash
API_URL=http://localhost:8000
```Build:
```bash
npm run build
```Run the server:
```bash
npm start
```## License
MIT License