10 lines
496 B
Markdown
Raw Normal View History

2024-02-05 06:51:58 -06:00
# Chapter 3: Coding Attention Mechanisms
2023-12-09 17:13:56 -06:00
## Main Chapter Code
2024-03-06 08:30:32 -06:00
- [01_main-chapter-code](01_main-chapter-code) contains the main chapter code.
## Bonus Materials
2024-07-26 08:45:36 -05:00
- [02_bonus_efficient-multihead-attention](02_bonus_efficient-multihead-attention) implements and compares different implementation variants of multihead-attention
- [03_understanding-buffers](03_understanding-buffers) explains the idea behind PyTorch buffers, which are used to implement the causal attention mechanism in chapter 3