mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2025-07-28 11:24:00 +00:00

* Use instance tokenizer * consistency updates --------- Co-authored-by: Sebastian Raschka <mail@sebastianraschka.com>
Chapter 3: Coding Attention Mechanisms
- 01_main-chapter-code contains the main chapter code.
- 02_bonus_efficient-multihead-attention implements and compares different implementation variants of multihead-attention