Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/kyegomez/tree-attention-torch
An implementation of Tree-Attention in PyTorch because it's in JAX for some reason
https://github.com/kyegomez/tree-attention-torch
ai attention dao deep-learning distributed gpu llm machine-learning ml nvidia parallel research stanford tri
Last synced: about 1 month ago
JSON representation
An implementation of Tree-Attention in PyTorch because it's in JAX for some reason
- Host: GitHub
- URL: https://github.com/kyegomez/tree-attention-torch
- Owner: kyegomez
- License: mit
- Created: 2024-09-09T19:25:26.000Z (5 months ago)
- Default Branch: main
- Last Pushed: 2024-11-04T05:30:28.000Z (3 months ago)
- Last Synced: 2024-11-09T08:41:18.405Z (3 months ago)
- Topics: ai, attention, dao, deep-learning, distributed, gpu, llm, machine-learning, ml, nvidia, parallel, research, stanford, tri
- Language: Python
- Homepage: https://discord.com/servers/agora-999382051935506503
- Size: 2.16 MB
- Stars: 4
- Watchers: 2
- Forks: 0
- Open Issues: 4
-
Metadata Files:
- Readme: README.md
- Funding: .github/FUNDING.yml
- License: LICENSE
Awesome Lists containing this project
README
[![Multi-Modality](agorabanner.png)](https://discord.com/servers/agora-999382051935506503)
# Tree Attention Torch
An implementation of Tree-Attention in PyTorch because it's in JAX for some reason
[![Join our Discord](https://img.shields.io/badge/Discord-Join%20our%20server-5865F2?style=for-the-badge&logo=discord&logoColor=white)](https://discord.gg/agora-999382051935506503) [![Subscribe on YouTube](https://img.shields.io/badge/YouTube-Subscribe-red?style=for-the-badge&logo=youtube&logoColor=white)](https://www.youtube.com/@kyegomez3242) [![Connect on LinkedIn](https://img.shields.io/badge/LinkedIn-Connect-blue?style=for-the-badge&logo=linkedin&logoColor=white)](https://www.linkedin.com/in/kye-g-38759a207/) [![Follow on X.com](https://img.shields.io/badge/X.com-Follow-1DA1F2?style=for-the-badge&logo=x&logoColor=white)](https://x.com/kyegomezb)
## Usage
```bash
python3 model.py
```# License
MIT# Todo
- [ ] Implement flash attention from the native official repo, I couldn't because the docs are nowhere to be found and understood