Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/ex3ndr/llama-coder

Replace Copilot local AI
https://github.com/ex3ndr/llama-coder

Last synced: 5 days ago
JSON representation

Replace Copilot local AI

Awesome Lists containing this project

README

        

# Llama Coder

Llama Coder is a better and self-hosted Github Copilot replacement for [VS Code](https://github.com/microsoft/vscode). Llama Coder uses [Ollama](https://ollama.ai) and codellama to provide autocomplete that runs on your hardware. Works best with Mac M1/M2/M3 or with RTX 4090.

[VS Code Plugin](https://marketplace.visualstudio.com/items?itemName=ex3ndr.llama-coder)

## Features
* 🚀 As good as Copilot
* ⚡️ Fast. Works well on consumer GPUs. Apple Silicon or RTX 4090 is recommended for best performance.
* 🔐 No telemetry or tracking
* 🔬 Works with any language coding or human one.

## Recommended hardware

Minimum required RAM: 16GB is a minimum, more is better since even smallest model takes 5GB of RAM.
The best way: dedicated machine with RTX 4090. Install [Ollama](https://ollama.ai) on this machine and configure endpoint in extension settings to offload to this machine.
Second best way: run on MacBook M1/M2/M3 with enough RAM (more == better, but 10gb extra would be enough).
For windows notebooks: it runs good with decent GPU, but dedicated machine with a good GPU is recommended. Perfect if you have a dedicated gaming PC.

## Local Installation

Install [Ollama](https://ollama.ai) on local machine and then launch the extension in VSCode, everything should work as it is.

## Remote Installation

Install [Ollama](https://ollama.ai) on dedicated machine and configure endpoint to it in extension settings. Ollama usually uses port 11434 and binds to `127.0.0.1`, to change it you should set `OLLAMA_HOST` to `0.0.0.0`.

## Models

Currently Llama Coder supports only Codellama. Model is quantized in different ways, but our tests shows that `q4` is an optimal way to run network. When selecting model the bigger the model is, it performs better. Always pick the model with the biggest size and the biggest possible quantization for your machine. Default one is `stable-code:3b-code-q4_0` and should work everywhere and outperforms most other models.

| Name | RAM/VRAM | Notes |
|---------------------------|----------|-------|
| stable-code:3b-code-q4_0 | 3GB | |
| codellama:7b-code-q4_K_M | 5GB | |
| codellama:7b-code-q6_K | 6GB | m |
| codellama:7b-code-fp16 | 14GB | g |
| codellama:13b-code-q4_K_M | 10GB | |
| codellama:13b-code-q6_K | 14GB | m |
| codellama:34b-code-q4_K_M | 24GB | |
| codellama:34b-code-q6_K | 32GB | m |

* m - slow on MacOS
* g - slow on older NVidia cards (pre 30xx)

## Troubleshooting

Most of the problems could be seen in output of a plugin in VS Code extension output.

## Changelog

## [0.0.14]
- Ability to pause completition (by @bkyle)
- Bearer token support for remote inference (by @Sinan-Karakaya)

## [0.0.13]
- Fix remote files support

## [0.0.12]
- Remote support
- Fix codellama prompt preparation
- Add trigger delay
- Add jupyter notebooks support

## [0.0.11]
- Added Stable Code model
- Pause download only for specific model instead of all models

## [0.0.10]
- Adding ability to pick a custom model
- Asking user if they want to download model if it is not available

## [0.0.9]
- Adding deepseek 1b model and making it default

## [0.0.8]
- Improved DeepSeek support and language detection

## [0.0.7]
- Added DeepSeek support
- Ability to change temperature and top p
- Fixed some bugs

## [0.0.6]
- Fix ollama links
- Added more models

## [0.0.4]

- Initial release of Llama Coder