Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/vs4vijay/ai-playground
Playground for AI, LLM, Agents
https://github.com/vs4vijay/ai-playground
agent ai anthropic autogen autogpt chatgpt cohere command-r gemma gpt-3 gpt-4 langchain llama llama2 llm mistral mixtral prompt semantic-kernel
Last synced: about 1 month ago
JSON representation
Playground for AI, LLM, Agents
- Host: GitHub
- URL: https://github.com/vs4vijay/ai-playground
- Owner: vs4vijay
- License: mit
- Created: 2023-06-13T18:36:41.000Z (over 1 year ago)
- Default Branch: main
- Last Pushed: 2024-09-23T17:10:27.000Z (3 months ago)
- Last Synced: 2024-11-08T21:05:15.875Z (about 1 month ago)
- Topics: agent, ai, anthropic, autogen, autogpt, chatgpt, cohere, command-r, gemma, gpt-3, gpt-4, langchain, llama, llama2, llm, mistral, mixtral, prompt, semantic-kernel
- Language: Python
- Homepage:
- Size: 106 KB
- Stars: 1
- Watchers: 3
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# AI-Playground
---
AI Playground for trying out LLM Models, Embeddings, Vector Stores, Semantic Search, RAG, Azure OpenAI, LLaMa, Mistral
## Installation
```bash
pip install -U ai-playground
```### Local Installation
Pre-requisites:
- Python 3.10+ and pip
```bash
# Start virtual environment
source ./activate# Install requirements
pip install -r requirements.txt
```### Running the full playground
- Copy `.env.example` to `.env` and fill in the values
- Run the following command to start the server
```bash
python ai_playground.py
```---
## Models
- Llama 2 - https://huggingface.co/TheBloke/Llama-2-7b-Chat-GGUF
- Llama 3 Instruct - https://huggingface.co/lmstudio-community/Meta-Llama-3-8B-Instruct-GGUF/tree/mainwget -c https://huggingface.co/TheBloke/Mistral-7B-v0.1-GGUF/resolve/main/mistral-7b-v0.1.Q8_0.gguf
wget -c https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/resolve/main/mistral-7b-instruct-v0.1.Q8_0.gguf
wget -c https://huggingface.co/TheBloke/Mistral-7B-OpenOrca-GGUF/resolve/main/mistral-7b-openorca.Q8_0.ggufwget -c https://huggingface.co/TheBloke/Wizard-Vicuna-7B-Uncensored-GGUF/resolve/main/Wizard-Vicuna-7B-Uncensored.Q8_0.gguf
wget -c https://huggingface.co/TheBloke/CodeLlama-7B-Instruct-GGUF/resolve/main/codellama-7b-instruct.Q8_0.gguf
---
# Run Local LLM
## Try with LLaMA.cpp
- Extract LLaMA.cpp zip to bin/ directory
```bash
./bin/main.exe -m models/llama-2-7b-chat.Q8_0.gguf
```## Try with ollama
```bash
echo "FROM ./models/llama-2-13b-chat.Q5_K_M.gguf" > llama-2-13b-chat.Modelfileollama create llama2-13b-chat -f ./llama-2-13b-chat.Modelfile
ollama run llama2-13b-chat
# OR
ollama run llama3.1:7b
ollama serve llama3.1:7bollama ps
curl -X POST http://localhost:11434/api/generate -d '{
"model": "wizardlm-uncensored",
"prompt":"Who made Rose promise that she would never let go?"
}'ollama run starcoder2:3b
ollama pull nomic-embed-text```
## Try with vLLM
```bash
pip install -U vllmpython -u -m vllm.entrypoints.openai.api_server --host 0.0.0.0 --model mistralai/Mistral-7B-v0.1
```## Try with FastChat
```bash
pip install -U fastchatpython -m fastchat.serve.openai_api_server --host localhost --port 8000
```## Try with LeptonAI
```bash
pip install -U leptonai```
## Use StarCoder
pip install transformers
pip install torch torchvision
pip install accelerate bitsandbytes
pip install accelerate[torch]Edit:
- `load_in_8bit=True`python starcoder.py
- will download ~60 GB of model---
## Specs
RAM Required:
| Model Size | RAM Required |
|------------|--------------|
| 3B | 8 GB |
| 7B | 16 GB |
| 13B | 32 GB |---
## Chat UIs
- OpenWebUI - https://github.com/open-webui/open-webui
- ChatBotUI - https://github.com/mckaywrigley/chatbot-ui
- OpenUI
- AnythingLLM
- LobeChat
- LibreChat - https://github.com/danny-avila/LibreChat
- oterm
- https://github.com/billmei/every-chatgpt-gui
- https://github.com/Bin-Huang/chatbox
- https://bettergpt.chat/---
## Code Assistants / Copilots
- Github Copilot
- Continue.dev - https://github.com/continuedev/continue
- Cursor
- LlamaCoder
- Zed AI
- ClaudeDev
- BlackBox AI
- Tabnine AI---
## Agents
- aider-chat - `pipx install aider-chat`
- metagpt - `pipx install metagpt`
- https://github.com/joaomdmoura/crewAI
----
## Other Tools
- https://github.com/outlines-dev/outlines
- guidance---
### Development Notes
```bash
pip install pyautogen
pip install openplayground
openplayground runollama run mistral
pip install -U jina
Ray Serve
pip install "ray[serve]"
https://github.com/ray-project/ray-llmtxtai
MLC AI - https://mlc.ai/package/
pip install --pre --force-reinstall mlc-ai-nightly mlc-chat-nightly -f https://mlc.ai/wheels
python -m mlc_chat.restOpenLLM
https://github.com/FlowiseAI/Flowise
wget https://gpt4all.io/models/ggml-gpt4all-j.bin -O models/ggml-gpt4all-j
https://github.com/go-skynet/LocalAI
docker pull quay.io/go-skynet/local-ai:latestnlpcloud
curl "https://api.nlpcloud.io/v1//entities" \
-H "Authorization: Token " \
-H "Content-Type: application/json" \
-X POST \
-d '{"text":"John Doe has been working for Microsoft in Seattle since 1999."}'https://github.com/microsoft/semantic-kernel
https://github.com/microsoft/guidancehttps://skypilot.readthedocs.io/
Later:
https://github.com/Arize-ai/phoenix
https://github.com/explodinggradients/ragas
https://github.com/trypromptly/LLMStackQ5_K_M
poetry export -f requirements.txt --output requirements.txt
poetry publish --buildlazypredict
mito
pip install langchain-serve
LangServe
pip install -U "langserve[all]"
pip install -U langchain-clilangflow run
flowise
promptflow
pip install promptflow promptflow-toolsPyGWalker
# DSPy
pip install dspy-aihttps://github.com/ShreyaR/guardrails
https://github.com/guardrails-ai/guardrailsguidance
https://github.com/1rgs/jsonformerLangChain
https://github.com/jina-ai/langchain-serveLangFlow / Flowise / LangSmith
ChainLitpromptflow
LMQI
https://github.com/eth-sri/lmqlhttps://github.com/zilliztech/GPTCache
https://github.com/argilla-io/argilla
https://github.com/vllm-project/vllm
https://github.com/TransformerOptimus/SuperAGI
accelerate
- accelerate config
- accelerate env
bitsandbytes
wand
https://github.com/huggingface/text-generation-inferencectransformers
spacy
spacy-llm
gorilla-cli
https://github.com/langgenius/dify
gptcachehttps://github.com/wandb/openui
```