Projects in Awesome Lists tagged with multihead-attention
A curated list of projects in awesome lists tagged with multihead-attention .
https://github.com/tlatkowski/multihead-siamese-nets
Implementation of Siamese Neural Networks built upon multihead attention mechanism for text semantic similarity task.
attention deep-architectures deep-learning deep-neural-networks multihead-attention multihead-attention-networks natural-language-processing nlp paraphrase paraphrase-identification python3 quora-question-pairs semantic-similarity sentence-similarity siamese-cnn siamese-lstm siamese-neural-network snli tensorflow text-similarity
Last synced: 13 Apr 2025
https://github.com/changwookjun/transformer
Chatbot using Tensorflow (Model is transformer) ko
bert chatbot korean-nlp multihead-attention self-attention tensorflow transformer
Last synced: 10 Apr 2025
https://github.com/abhilash1910/graphattentionnetworks
This package is a Tensorflow2/Keras implementation for Graph Attention Network embeddings and also provides a Trainable layer for Multihead Graph Attention.
graph-attention-networks keras-tensorflow leaky-relu multihead-attention self-attention tf2
Last synced: 10 Apr 2025
https://github.com/Resh-97/MixSeq-Connecting-Macroscopic-Time-Series-Forecasting-with-Microscopic-Time-Series-Data
Testing the Reproducibility of the paper: MixSeq. Under the assumption that macroscopic time series follow a mixture distribution, they hypothesise that lower variance of constituting latent mixture components could improve the estimation of macroscopic time series.
arma comp6248 deepar multihead-attention reproducibility-challenge time-series vae-implementation vae-pytorch
Last synced: 26 Mar 2025
https://github.com/dcarpintero/transformer101
Annotated vanilla implementation in PyTorch of the Transformer model introduced in 'Attention Is All You Need'.
attention-is-all-you-need dot-product-attention dropout-layers encoder-decoder-architecture feedforward-neural-network gelu linear-layers multihead-attention normalization-layers positional-encoding pytorch self-attention softmax transfomer
Last synced: 14 Mar 2025
https://github.com/aniketdash7/multihead_attention_implementation
Implementation of Multihead attention mechanism using numpy and pyTorch
multihead-attention numpy pytorch torch
Last synced: 08 Feb 2025
https://github.com/jivanacharya/shakespeare-gpt
Implementing a GPT (Generative Pre-trained Transformer) model from scratch on Shakespeare's work.
gpt multihead-attention self-attention transformer
Last synced: 03 Apr 2025
https://github.com/aman-17/3dprinting-extrusion-detection
3D Printing Extrusion Detection using Multi-Head Attention Model
3d-printing deep-learning multihead-attention python
Last synced: 20 Mar 2025