An open API service indexing awesome lists of open source software.

https://github.com/mydarapy/smollm-experiments-with-grouped-query-attention

(Unofficial) building Hugging Face SmolLM-blazingly fast and small language model with PyTorch implementation of grouped query attention (GQA)
https://github.com/mydarapy/smollm-experiments-with-grouped-query-attention

attention grouped-query-attention huggingface huggingface-smol-lm llm ml-efficiency smol smol-lm transformer

Last synced: about 1 month ago
JSON representation

(Unofficial) building Hugging Face SmolLM-blazingly fast and small language model with PyTorch implementation of grouped query attention (GQA)

Awesome Lists containing this project

README

        

### Improving Sub-billion Scale LLM Design Experiments
Some of the techniques used in the LLM pretraining design include:
- Embedding Sharing
- Grouped Query Design
- SwiGLU Activations for the Multi Perceptron Layer
- Intermidate blockwise weight sharing

chrome_7KTClIuUnl