Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/jonboh/shai
A Shell AI assistant. It can help you write and explain commands, without having to leave your comfy terminal!
https://github.com/jonboh/shai
ai chatgpt openassistant rust shell terminal
Last synced: 3 days ago
JSON representation
A Shell AI assistant. It can help you write and explain commands, without having to leave your comfy terminal!
- Host: GitHub
- URL: https://github.com/jonboh/shai
- Owner: jonboh
- License: mit
- Created: 2023-07-15T15:08:06.000Z (over 1 year ago)
- Default Branch: main
- Last Pushed: 2024-01-14T10:06:15.000Z (10 months ago)
- Last Synced: 2024-04-14T12:10:30.059Z (7 months ago)
- Topics: ai, chatgpt, openassistant, rust, shell, terminal
- Language: Rust
- Homepage: https://jonboh.dev/posts/shai/
- Size: 44 MB
- Stars: 13
- Watchers: 2
- Forks: 1
- Open Issues: 1
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Shai. Shell AI Assitant
Shai is a Shell AI assistant. Its purpose is to help you interact with your machine through
a command line interface. It has two main ways to achieve this.
- Command Generation: Shai allows you to quickly ask a LLM model to generate a command based on what it should do.
- Command Explanation: Given a command Shai will explain what a command will do and its side effects.You'll probably interact with Shai primarily in its command generation mode.
![main example](./assets/git.gif)You can also ask Shai to explain the command it just generated:
![explain generated command](./assets/explanation.gif)or a command that is currently in your buffer line:
![explain buffer line command](./assets/explanation_buffer.gif)Note that the explanation will not take into account your initial prompt. This is purposely done to avoid biasing
the model explanation. In addition this can help double check that the model generated the correct command and detect
model hallucinations.The model explanations will briefly explain the command and enumerate a list of common side-effects.
In addition, if the command is potentially dangerous the model will usually point it out.As a backend you can use any of the following OpenAI models:
- GPT-3.5-turbo
- GPT-3.5-turbo-16k
- GPT-4
- GPT-4-32kThis might be anecdotal, but I've sometimes experienced lower latencies with the larger context models.
For the time being I don't have access to the gpt-4 models API, but they should work if your account
does have access. All examples have been generated with GPT-3.5-turbo, so I expect the results from
GPT-4 to be equivalent or better (specially when the request has several moving parts, like pipelines
or redirection).Support for other models is planned. See [Current Status](#current-status)
Note that Shai is not intended to replace a solid knowledge of your system or its commands, but it
can help ease the burden of memorizing every single flag and command syntax. The OpenAI GPT models are
by no means foolproof, and the best results are usually achieved when you as an operator already know
the terminology and capabilities related to the command you want to generate. It can also be a resource
to explain commands taken from tutorials or forums.## Installation
##### Cargo
```bash
cargo install shai
```
Remember to add `.cargo/bin` to your `PATH`.This method should be available for all platforms, but you'll need to have Cargo installed, use
[rustup](https://rustup.rs/) to install it.Note: See the [next section](#shell-integration) to integrate shai and allow it
to interact with your buffer line.##### Arch Linux. AUR:
```bash
yay -S shai
```##### Other
Use the [latest release](https://github.com/jonboh/shai/releases/latest) in this github repo.
Put the binary in a location that is included in your `PATH`, source the corresponding integration
shell script (see [next section](#shell-integration)) in your rc file and you should be good to go.### Shell Integration
I've developed Shai with the aim to have it very closely integrated in the shell. I tend to use it similarly to
[fzf](https://github.com/junegunn/fzf), so one key-shortcut away.In this repository you will find the scripts needed to integrate Shai in your shell experience.
The following list contains the shells that are currently supported and the planned ones.- [x] Bash
- [x] Zsh
- [x] Fish
- [x] Nushell
- [x] PowerShell
- [ ] Window CMDIn order for Shai to be available through shortcuts you'll need to source the corresponding
script. These scripts provide a default binding, modify it to adapt it to your needs.
All these scripts perform the same function, the setup a function that is called through the keybind.
The function takes the current content of your shell buffer, forwards it to Shai, and if the
command generation interface is being used and Shai is exited with `Ctrl+a` (accept), the buffer
line is modified with the command generated. In most cases the function essentially highjacks the
command edition capabilities of the shell treating Shai as a text editor.| Shell | Integration file |
| --------------- | --------------- |
| Bash | `bash_assistant.sh` |
| Zsh | `zsh_assistant.zsh` |
| Fish | `fish_assistant.fish` |
| Nushell | `nushell_assistant.nu` |
| PowerShell | `powershell_assistant.ps1` |If you just install the binary you can generate the integration script using shai
```bash
shai generate-script --shell
# e.g
shai generate-script --shell zsh > zsh_assistant.zsh
# then in your .zshrc
source zsh_assistant.zsh
```
Remember to source the resulting script in you rc, otherwise you won't have the shortcuts available.Depending on what model you use you might need to provide the API key as an environment variable. For the OpenAI models
you could set it with:
```bash
export OPENAI_API_KEY=$()
# if you have it on a text file
export OPENAI_API_KEY=$(cat ~/.secrets/chatgpt.key)
```
See [here](https://help.openai.com/en/articles/4936850-where-do-i-find-my-secret-api-key) how to get your OpenAI API key.## How to use it
### Shell Shortcuts
If you don't modify the shell integration script the shortcuts will be the following:
- `Alt+s` : Command Generation
- `Alt+e` : Command ExplanationYou can change these shortcuts on the shell integration files. They have been chosen to avoid collision with already existing
shortcuts in *readline*.### Shai Controls
- `Enter` : Send Prompt
- `Ctrl+c` : Exit Shai
- `Esc` : Cancel current request#### Command Generation
When a command is generated the following controls are also available:
- `Ctrl+a` : Accept generated command
- This will check that the model followed the format instructions, when the model provides a long answer
this shortcut will retrieve only the text inside markdown code blocks.
- `Ctrl+r` : Accept generated command (raw)
- This will accept the input from the model without any check
- `Ctrl+e` : Explain generated command
- This will generate an explanation for the command that Shai has just generated.These keybinds cannot currently be changed.
## Examples
### Git
![Get the commit hash in which a string was introduced](./assets/git_log_insert_string.png)
![Get the commit hash of the commits that include string](./assets/git_message_grep.png)
![Get the commit hash of the commits that modify a file](./assets/git_hashes_file.png)### Miscellaneous
Run bash on a running container:
![run bash on a running container](./assets/docker_exec.gif)Make a ssh tunnel:
![SSH Tunnel](./assets/ssh_tunnel.png)Simple ffmpeg operations, or command modifications:
![Make a GIF from a video file](./assets/ffmpeg_gif.png)
![Modify a ffmpeg command](./assets/modify_ffmpeg.png)Discard command output:
![Discard output of a terminal](./assets/discard_output.gif)## Current Status
At the moment Shai is memoryless, when you send a prompt it is sent to the model without any
context from your previous prompts. The purpose of Shai is not to be a conversational application but
to quickly generate commands leveraging LLMs. I might implement conversational functionality later in
the future if it improves the commands produced.I plan to add support for other models like [OpenAssistant](https://github.com/LAION-AI/Open-Assistant)
once they provide an API or a way to run it locally is available (even if the requirements for the machine
are steep, but within a high-end desktop). If you have pointers for other similar models that can be run
locally or have an available API please, fill an issue and I will try to add support for it.### Context Awareness.
You can modify the assumed Operating System or distro to get more relevant results. This information
is provided in the `--operating-system` option. You can modify your integration script with the appropriate
value.In the same way the `--shell` option lets the model know in which shell it is running, this can help the model
use shell specific features. However, for modern shells like `nushell` this can actually confuse the model as it won't
have a lot of information about this shell in its training data. In cases like this it might be better to
let the model believe it is running on another shell (like `bash`).#### Experimental Options
Initially I envisioned Shai as a more capable assistant to which context about the current state
of your machine could be forwarded, and it would act accordingly. There are some options that are
disabled by default:| Option | Description |
| ---------------- | --------------- |
| `pwd` | Provides the model with the current working directory |
| `depth` | depth with which to run the `tree` command. It provides context about ther current directory and its content |
| `environment` | The list of environment variables set (only their name is passed to the model) |
| `programs` | The list of available programs to the model with which to complete the task |I have found that the performance of the GPT3.5 model is lacking in this respect. I have
some hope that the GPT-4 (or future) models perform better, so the option to forward this context
to the model is still available, if you find good results with it let me know!## Roadmap
In no particular order:
- [ ] Custom keybindings
- [ ] Prompt Overrides
- [ ] Add prompt history navigation.
- [ ] Add support for other models. (I'd like to make it easy to interact with a locally run model).
- [x] OpenAI
- [ ] Local
- [ ] OpenAssistant (once a public API is available)
- [ ] Google Bard (once a public API is available)
- [ ] Add other shells
- [x] Bash
- [x] Zsh
- [x] Fish
- [x] Nushell
- [x] PowerShell
- [ ] Window CMD## Acknowlegments
- [fzf](https://github.com/junegunn/fzf)
- fzf is the main inspiration in terms of the integration with the shell
- [ChatGPT-rs](https://github.com/Maxuss/chatgpt_rs)
- I did take some hints from this project to implement the streaming responses from OpenAI.