https://github.com/keshavpatel2/local-llm-workbench
🧠A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed guides to maximize LLM performance on your hardware.
https://github.com/keshavpatel2/local-llm-workbench
context-window-scaling cpu-inference cuda gpu-acceleration hybrid-inference inference-optimization llama-cpp llm-benchmarking llm-deployment local-llm model-management model-quantization ollama-optimization wsl-ai-setup
Last synced: 3 months ago
JSON representation
🧠A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed guides to maximize LLM performance on your hardware.
- Host: GitHub
- URL: https://github.com/keshavpatel2/local-llm-workbench
- Owner: Keshavpatel2
- Created: 2025-03-30T12:21:45.000Z (3 months ago)
- Default Branch: main
- Last Pushed: 2025-03-30T13:29:17.000Z (3 months ago)
- Last Synced: 2025-03-30T13:34:57.716Z (3 months ago)
- Topics: context-window-scaling, cpu-inference, cuda, gpu-acceleration, hybrid-inference, inference-optimization, llama-cpp, llm-benchmarking, llm-deployment, local-llm, model-management, model-quantization, ollama-optimization, wsl-ai-setup
- Language: Shell
- Size: 8.79 KB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0