Projects in Awesome Lists tagged with cpu-inference
A curated list of projects in awesome lists tagged with cpu-inference .
https://github.com/kennethleungty/Llama-2-Open-Source-LLM-CPU-Inference
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
c-transformers chatgpt cpu cpu-inference deep-learning document-qa faiss langchain language-models large-language-models llama llama-2 llm machine-learning natural-language-processing nlp open-source-llm python sentence-transformers transformers
Last synced: 09 May 2025
https://github.com/kennethleungty/llama-2-open-source-llm-cpu-inference
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
c-transformers chatgpt cpu cpu-inference deep-learning document-qa faiss langchain language-models large-language-models llama llama-2 llm machine-learning natural-language-processing nlp open-source-llm python sentence-transformers transformers
Last synced: 16 May 2025
https://github.com/rbitr/llm.f90
LLM inference in Fortran
ai chatbot cpu-inference language-model llama llama2 llamacpp llm mamba phi-2 state-space-model transformer
Last synced: 24 Jan 2025
https://github.com/jozsefszalma/homelab
The bare metal in my basement
ai bare-metal cpu-inference deep-learning gpu hardware-hacking hobby-project homelab machine-learning server
Last synced: 12 Mar 2025
https://github.com/yybit/pllm
Portable LLM - A rust library for LLM inference
Last synced: 10 Apr 2025
https://github.com/codito/arey
Simple large language model playground app
ai cli cpu-inference gguf large-language-models llama2 llamacpp llm local-model mistral ollama
Last synced: 10 Feb 2025
https://github.com/johnclaw/chatllm.v
V-lang api wrapper for llm-inference chatllm.cpp
api-wrapper bindings chatbot chatllm cpu-inference gemma ggml inference llama llm llm-inference llms mistral phi3 quantization qwen v-lang vlang
Last synced: 15 Apr 2025
https://github.com/laelhalawani/gguf_llama
Wrapper for simplified use of Llama2 GGUF quantized models.
cpu-inference gguf llama llama2 llamacpp quantization
Last synced: 24 Jan 2025
https://github.com/johnclaw/chatllm.vb
VB.NET api wrapper for llm-inference chatllm.cpp
api-wrapper bindings chatllm cpu-inference gemma ggml int8 int8-inference int8-quantization llama llm-inference mistral qwen vb-net vbnet
Last synced: 15 Apr 2025
https://github.com/johnclaw/chatllm.nim
Nim api-wrapper for llm-inference chatllm.cpp
api-wrapper bindings chatbot chatllm cpu-inference gemma ggml inference llama llm llm-inference llms mistral nim nim-lang nim-language nimlang phi quantization qwen
Last synced: 15 Apr 2025
https://github.com/johnclaw/chatllm.cs
C# api wrapper for llm-inference chatllm.cpp
api-wrapper bindings chatllm cpu-inference csharp gemma ggml inference int8 int8-inference int8-quantization llama llm llm-inference llms mistral qwen
Last synced: 15 Apr 2025
https://github.com/chinese-soup/cbot-telegram-whisper
Simple bot that transcribes Telegram voice messages. Powered by go-telegram-bot-api & whisper.cpp Go bindings.
bot cpu-inference golang openai speech-recognition speech-to-text whisper whisper-cpp whispercpp
Last synced: 10 Mar 2025
https://github.com/johnclaw/chatllm.rs
rust api wrapper for llm-inference chatllm.cpp
api-wrapper bindings chatbot chatllm cpu-inference gemma ggml inference llama llm llm-inference llms mistral quantization qwen rust
Last synced: 21 Mar 2025
https://github.com/johnclaw/chatllm.kt
kotlin api wrapper for llm-inference chatllm.cpp
api-wrapper bindings chatbot chatllm cpu-inference gemma ggml inference kotlin llama llm llm-inference llms mistral quantization qwen
Last synced: 20 Mar 2025
https://github.com/johnclaw/chatllm.lua
lua api wrapper for llm-inference chatllm.cpp
api-wrapper bindings chatbot chatllm cpu-inference gemma ggml inference llama llm llm-inference llms lua luajit mistral quantization qwen
Last synced: 18 Mar 2025
https://github.com/johnclaw/chatllm.d
D-lang api wrapper for llm-inference chatllm.cpp
api-wrapper bindings chatbot chatllm cpu-inference d-lang d-language dlang gemma ggml inference llama llm llm-inference llms mistral quantization qwen
Last synced: 21 Mar 2025
https://github.com/keshavpatel2/local-llm-workbench
🧠 A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed guides to maximize LLM performance on your hardware.
context-window-scaling cpu-inference cuda gpu-acceleration hybrid-inference inference-optimization llama-cpp llm-benchmarking llm-deployment local-llm model-management model-quantization ollama-optimization wsl-ai-setup
Last synced: 01 Apr 2025