https://github.com/nexusgpu/tensor-fusion
Tensor Fusion is a state-of-the-art GPU virtualization and pooling solution designed to optimize GPU cluster utilization to its fullest potential.
https://github.com/nexusgpu/tensor-fusion
ai amd-gpu autoscaling gpu gpu-pooling gpu-scheduling gpu-usage gpu-virtualization inference karpenter llm-serving nvidia pytorch rcuda remote-gpu vgpu
Last synced: 2 months ago
JSON representation
Tensor Fusion is a state-of-the-art GPU virtualization and pooling solution designed to optimize GPU cluster utilization to its fullest potential.
- Host: GitHub
- URL: https://github.com/nexusgpu/tensor-fusion
- Owner: NexusGPU
- License: apache-2.0
- Created: 2024-11-12T23:49:48.000Z (11 months ago)
- Default Branch: main
- Last Pushed: 2025-07-20T04:21:16.000Z (3 months ago)
- Last Synced: 2025-07-20T06:36:27.086Z (3 months ago)
- Topics: ai, amd-gpu, autoscaling, gpu, gpu-pooling, gpu-scheduling, gpu-usage, gpu-virtualization, inference, karpenter, llm-serving, nvidia, pytorch, rcuda, remote-gpu, vgpu
- Language: Go
- Homepage: https://tensor-fusion.ai
- Size: 1.51 MB
- Stars: 55
- Watchers: 2
- Forks: 13
- Open Issues: 5
-
Metadata Files:
- Readme: README.md
- License: LICENSE
- Security: SECURITY.md
Awesome Lists containing this project
README
TensorFusion.AI
Next-Generation GPU Virtualization and Pooling for Enterprises
Less GPUs, More AI Apps.
Explore the docs ยป
View Demo
|
Report Bug
|
Request Feature
# โพ๏ธ Tensor Fusion
[![Contributors][contributors-shield]][contributors-url]
[![Forks][forks-shield]][forks-url]
[![Stargazers][stars-shield]][stars-url]
[![Issues][issues-shield]][issues-url]
[![MIT License][license-shield]][license-url]
[![LinkedIn][linkedin-shield]][linkedin-url]
[](https://deepwiki.com/NexusGPU/tensor-fusion)Tensor Fusion is a state-of-the-art **GPU virtualization and pooling solution** designed to optimize GPU cluster utilization to its fullest potential.
## ๐ Highlights
#### ๐ Fractional GPU with Single TFlops/MiB Precision
#### ๐ Battle-tested GPU-over-IP Remote GPU Sharing
#### โ๏ธ GPU-first Scheduling and Auto-scaling
#### ๐ Computing Oversubscription and GPU VRAM Expansion
#### ๐ซ GPU Pooling, Monitoring, Live Migration, AI Model Preloading and more## ๐ฌ Demo
### Fractional vGPU & GPU-over-IP & Distributed Allocation

### AI Infra Console

### GPU Live-migration [End-to-end feature WIP]
https://cdn.tensor-fusion.ai/GPU_Content_Migration.mp4
## ๐ Quick Start
### Onboard Your Own AI Infra
- [Deploy in Kubernetes cluster](https://tensor-fusion.ai/guide/getting-started/deployment-k8s)
- [Create new cluster in VM/BareMetal](https://tensor-fusion.ai/guide/getting-started/deployment-vm)
- [Learn Essential Concepts & Architecture](https://tensor-fusion.ai/guide/getting-started/architecture)### ๐ฌ Discussion
- Discord channel: [https://discord.gg/2bybv9yQNk](https://discord.gg/2bybv9yQNk)
- Discuss anything about TensorFusion: [Github Discussions](https://github.com/NexusGPU/tensor-fusion/discussions)
- Contact us with WeCom for Greater China region: [ไผไธๅพฎไฟก](https://work.weixin.qq.com/ca/cawcde42751d9f6a29)
- Email us: [support@tensor-fusion.com](mailto:support@tensor-fusion.com)
- Schedule [1:1 meeting with TensorFusion founders](https://tensor-fusion.ai/book-demo)## ๐ฎ Features & Roadmap
### Core GPU Virtualization Features
- [x] Fractional GPU and flexible oversubscription
- [x] Remote GPU sharing with SOTA GPU-over-IP technology, less than 4% performance loss
- [x] GPU VRAM expansion and hot/warm/cold tiering
- [ ] None NVIDIA GPU/NPU vendor support### Pooling & Scheduling & Management
- [x] GPU/NPU pool management in Kubernetes
- [x] GPU-first scheduling and allocation, with single TFlops/MB precision
- [x] GPU node auto provisioning/termination
- [x] GPU compaction/bin-packing
- [x] Seamless onboarding experience for Pytorch, TensorFlow, llama.cpp, vLLM, Tensor-RT, SGlang and all popular AI training/serving frameworks
- [x] Centralized Dashboard & Control Plane
- [ ] GPU-first autoscaling policies, auto set requests/limits/replicas
- [ ] Request multiple vGPUs with group scheduling for large models
- [ ] Support different QoS levels### Enterprise Features
- [x] GPU live-migration, snapshot/distribute/restore GPU context cross cluster, fastest in the world
- [ ] AI model registry and preloading, build your own private MaaS(Model-as-a-Service)
- [ ] Advanced auto-scaling policies, scale to zero, rebalance of hot GPUs
- [ ] Advanced observability features, detailed metrics & tracing/profiling of CUDA calls
- [ ] Monetize your GPU cluster by multi-tenancy usage measurement & billing report
- [ ] Enterprise level high availability and resilience, support topology aware scheduling, GPU node auto failover etc.
- [ ] Enterprise level security, complete on-premise deployment support, encryption in-transit & at-rest
- [ ] Enterprise level compliance, SSO/SAML support, advanced audit, ReBAC control, SOC2 and other compliance reports available### ๐ณ๏ธ Platform Support
- [x] Run on Linux Kubernetes clusters
- [x] Run on Linux VMs or Bare Metal (one-click onboarding to Edge K3S)
- [x] Run on Windows (Docs not ready, contact us for support)
- [ ] Run on MacOS (Imagining mount a virtual NVIDIA GPU device on MacOS!)See the [open issues](https://github.com/NexusGPU/tensor-fusion/issues) for a full list of proposed features (and known issues).
## ๐ Contributing
Contributions are what make the open source community such an amazing place to learn, inspire, and create. Any contributions you make are **greatly appreciated**.
If you have a suggestion that would make this better, please fork the repo and create a pull request. You can also simply open an issue with the tag "enhancement".
Don't forget to give the project a star! Thanks again!1. Fork the Project
2. Create your Feature Branch (`git checkout -b feature/AmazingFeature`)
3. Commit your Changes (`git commit -m 'Add some AmazingFeature'`)
4. Push to the Branch (`git push origin feature/AmazingFeature`)
5. Open a Pull Request### Top contributors
## ๐ท License
1. This repo is open sourced with [Apache 2.0 License](./LICENSE), which includes **GPU pooling, scheduling, management features**, you can use it for free and modify it.
2. **GPU virtualization and GPU-over-IP features** are also free to use as the part of **Community Plan**, the implementation is not fully open sourced
3. Features mentioned in "**Enterprise Features**" above are paid, **licensed users can automatically unlock these features**.[](https://app.fossa.com/projects/git%2Bgithub.com%2FNexusGPU%2Ftensor-fusion?ref=badge_large&issueType=license)
[contributors-shield]: https://img.shields.io/github/contributors/NexusGPU/tensor-fusion.svg?style=for-the-badge
[contributors-url]: https://github.com/NexusGPU/tensor-fusion/graphs/contributors
[forks-shield]: https://img.shields.io/github/forks/NexusGPU/tensor-fusion.svg?style=for-the-badge
[forks-url]: https://github.com/NexusGPU/tensor-fusion/network/members
[stars-shield]: https://img.shields.io/github/stars/NexusGPU/tensor-fusion.svg?style=for-the-badge
[stars-url]: https://github.com/NexusGPU/tensor-fusion/stargazers
[issues-shield]: https://img.shields.io/github/issues/NexusGPU/tensor-fusion.svg?style=for-the-badge
[issues-url]: https://github.com/NexusGPU/tensor-fusion/issues
[license-shield]: https://img.shields.io/github/license/NexusGPU/tensor-fusion.svg?style=for-the-badge
[license-url]: https://github.com/NexusGPU/tensor-fusion/blob/master/LICENSE
[linkedin-shield]: https://img.shields.io/badge/-LinkedIn-black.svg?style=for-the-badge&logo=linkedin&colorB=555
[linkedin-url]: https://www.linkedin.com/company/tensor-fusion/about