The notebook attention-in-transformers.ipynb
implements the attention mechanisms covered in Josh Starmer's (StatQuest) DeepLearning.AI lesson, Attention in Transformers: Concepts and Code in PyTorch.
Topics covered include:
- Positional encodings.
- Self-attention.
- Masked self-attention.
- Encoder-decoder attention.
- Multi-head attention.