https://github.com/LINs-lab/DynMoE
[ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models
https://github.com/LINs-lab/DynMoE
adaptive-computation language-model mixture-of-experts moe multimodal-large-language-models
Last synced: 5 months ago
JSON representation
[ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models
- Host: GitHub
- URL: https://github.com/LINs-lab/DynMoE
- Owner: LINs-lab
- License: apache-2.0
- Created: 2024-05-17T08:25:31.000Z (over 1 year ago)
- Default Branch: main
- Last Pushed: 2025-02-07T14:54:17.000Z (8 months ago)
- Last Synced: 2025-02-07T15:36:39.980Z (8 months ago)
- Topics: adaptive-computation, language-model, mixture-of-experts, moe, multimodal-large-language-models
- Language: Python
- Homepage: https://arxiv.org/abs/2405.14297
- Size: 57.3 MB
- Stars: 67
- Watchers: 11
- Forks: 9
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- awesome-mixture-of-experts - 23 May 2024