Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/harrisonhemstreet/ask-ollama
Ask-Ollama is a command-line tool that allows users to interact with the Ollama models directly from the terminal. This tool provides a simple and intuitive way to ask questions and receive responses from Ollama models.
https://github.com/harrisonhemstreet/ask-ollama
ai chatbot cli llm ollama rust terminal
Last synced: about 1 month ago
JSON representation
Ask-Ollama is a command-line tool that allows users to interact with the Ollama models directly from the terminal. This tool provides a simple and intuitive way to ask questions and receive responses from Ollama models.
- Host: GitHub
- URL: https://github.com/harrisonhemstreet/ask-ollama
- Owner: HarrisonHemstreet
- Created: 2023-11-30T23:39:06.000Z (about 1 year ago)
- Default Branch: main
- Last Pushed: 2023-12-01T04:53:10.000Z (about 1 year ago)
- Last Synced: 2024-11-20T17:49:26.297Z (about 2 months ago)
- Topics: ai, chatbot, cli, llm, ollama, rust, terminal
- Language: Rust
- Homepage: https://crates.io/crates/ask-ollama
- Size: 18.6 KB
- Stars: 13
- Watchers: 2
- Forks: 0
- Open Issues: 5
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Ask-Ollama
Ask-Ollama is a command-line tool that allows users to interact with the [Ollama](https://ollama.ai/) LLM models directly from the terminal. This tool provides a simple and intuitive way to ask questions and receive responses from Ollama models.
## Features
- **Interactive CLI**: Easily ask questions and get responses.
- **Model Selection**: Choose different models for varied responses.## Installation
To install Ask-Ollama, you need to have Rust and Cargo installed on your system. If you haven't already installed Rust, you can do so by following the instructions [here](https://www.rust-lang.org/tools/install).
Once Rust is installed, you can install Ask-Ollama using Cargo:
```sh
cargo install ask-ollama
```## Usage
After installation, you can start using Ask-Ollama by running:
```sh
ask [OPTIONS] [PROMPT]
```### Options
- `--model=[MODEL]`: Specify the model to use (default is 'mistral').
- `--version`: Display the version of the installed Ask-Ollama tool.
- `[PROMPT]`: The question or prompt to send to Ollama. Quotation marks are not required.### Examples
Asking a question using the default model:
```sh
ask "What is the capital of France?"
```
or
```sh
ask What is the capital of France?
```Specifying a different model:
```sh
ask --model=gale "Explain the theory of relativity"
```
Find all available models from Ollama [here](https://ollama.ai/library).Checking the version:
```sh
ask --version
```Seeing the help info:
```sh
ask --help
```### Watch the YouTube Video Here! (click on the image)
[![How to use ask-ollama](https://i.imgur.com/Pgr5FZl.png)](https://www.youtube.com/watch?v=05GynRi9NAQ)## Contributing
Contributions to Ask-Ollama are welcome! If you have suggestions for improvements or encounter any issues, please feel free to open an issue or submit a pull request on our [GitHub repository](https://github.com/HarrisonHemstreet/ask-ollama).
## License
Ask-Ollama is licensed under the MIT License.