An open API service indexing awesome lists of open source software.

https://github.com/ostad-ai/large-language-models

THis repository includes topics related to the Large Language Models (LLMs)
https://github.com/ostad-ai/large-language-models

large-language-models llm self-attention

Last synced: 6 months ago
JSON representation

THis repository includes topics related to the Large Language Models (LLMs)

Awesome Lists containing this project

README

          

# Large Language Models (LLMs)
(under construction)
1) What is a Large Language Model (LMM)?
2) What is the building block of an LLM?
3) **LLMs, self-attention mechanism:** The self-attention mechanism is the core concept of **transformer**-based LLMs. Here, we review the formulae of this mechanism and implement a self-attention from scratch in Python.
4) **LLMs, the softmax in self-attention:** We remind the softmax function ,which is widely used in *neural networks*, *deep learning*, and *machine learning*. The function softmax is implemented in Python with an example.
5) **LLMs: Layer normalization:** Layer normalization is a critical component of *Transformers* and *LLMs*, ensuring stable and efficient training by normalizing activations across the *feature dimension*. It is particularly well-suited for sequence-based tasks and deep architectures. Here, we implement the layer normalization with Numpy. Moreover, we give the code of *PyTorch* for the layer normalization so that you can compare the results.