Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/richawo/minimal-llm-ui
Minimalistic UI for Ollama LMs - This powerful react interface for LLMs drastically improves the chatbot experience and works offline.
https://github.com/richawo/minimal-llm-ui
ollama ollama-gui ollama-interface ollama-ui
Last synced: about 1 month ago
JSON representation
Minimalistic UI for Ollama LMs - This powerful react interface for LLMs drastically improves the chatbot experience and works offline.
- Host: GitHub
- URL: https://github.com/richawo/minimal-llm-ui
- Owner: richawo
- License: mit
- Created: 2023-10-17T21:53:55.000Z (about 1 year ago)
- Default Branch: main
- Last Pushed: 2024-09-21T08:05:14.000Z (about 2 months ago)
- Last Synced: 2024-09-30T05:27:01.674Z (about 1 month ago)
- Topics: ollama, ollama-gui, ollama-interface, ollama-ui
- Language: TypeScript
- Homepage:
- Size: 3.21 MB
- Stars: 250
- Watchers: 5
- Forks: 43
- Open Issues: 4
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- Awesome-Ollama - Minimalistic React UI for Ollama Models
README
Minimalistic Interface for Local Language Models (LLMs) (Powered by Ollama)
=========================================================Introduction
------------[![Watch the video](assets/preview.png)](https://github.com/richawo/minimal-llm-ui/assets/35015261/666053ed-cba8-45a6-9a96-ff8b2b525dbd)
This minimalistic UI is designed to act as a simple interface for Ollama models, allowing you to chat with your models, save conversations and toggle between different ones easily. The tool is built using React, Next.js, and Tailwind CSS, with LangchainJs and Ollama providing the magic behind the scenes.
Features
--------* Chat with Local Language Models (LLMs): Interact with your LLMs in real-time through our user-friendly interface.
* Model Toggling: You can switch between different LLMs easily (even mid-conversation), allowing you to experiment and explore different models for various tasks.
* Memory-based Context Storage: Keep track of context in memory, ensuring smooth interactions even when switching between models.
* Conversation History: Save conversations in a local database, allowing you to revisit them later.
* Prompt Templating: Save prompts you love by creating parameter-driven prompt templates to improve reuse
* Custom API Endpoint: Configure a custom base URL easily if Ollama is running on a different host/device.Technical Details
----------------* Built using React, Next.js, and Tailwind CSS for a clean and modern design.
* Utilizes LangchainJs and Ollama for seamless integration with Local Language Models (LLMs).
* Stores context in memory for efficient model switching.Getting Started
---------------1. Download and run Ollama on your machine with `ollama serve` or `ollama run ` (it will run at: http://localhost:11434/)
2. Open a new terminal and navigate to the root of this project.
3. Install the dependencies `npm install` in your terminal.4. Also check whether your node by doing:
```
node -v
```
If it is less than 14.0.1.
You can do this to update it:
- Install n using npm (Node.js package manager):- bash:
```
npm install -g n
```
Use n to install a specific Node.js version:
bash:
```
n 20.0.9
```
Verify the Node.js version:
- bash
```
node -v
```5. **Optional:** If running Ollama on a different host/device, customize the Ollama API base URL by copying `.env.example` to `.env.local` and setting the environment variable `NEXT_PUBLIC_OLLAMA_BASEURL`. If not set, the base URL will default to `http://localhost:11434`.
6. Start the tool by running `npm run dev` (it should be available in your web browser at `http://localhost:3000`)To-do
---------------- Add edit message icon for user messages
- Add image uploads for multi-modal models
- Summarise conversations
- Incorporate visualisations
- Convert to a desktop app so that it can be more powerful
- Command menu should let you edit + delete existing promptsTroubleshooting
---------------If you encounter any issues, feel free to reach out!
License
-------This project is licensed under the MIT License. See `LICENSE` file for details.