mirror of
				https://github.com/rasbt/LLMs-from-scratch.git
				synced 2025-10-31 09:50:23 +00:00 
			
		
		
		
	 cba4f89514
			
		
	
	
		cba4f89514
		
	
	
	
	
		
			
			* updated Dockerfile * updated MHA implementations for PT 2.5 * fixed typo * update installation instruction * Update setup/03_optional-docker-environment/.devcontainer/Dockerfile --------- Co-authored-by: rasbt <mail@sebastianraschka.com>
Chapter 3: Coding Attention Mechanisms
Main Chapter Code
- 01_main-chapter-code contains the main chapter code.
Bonus Materials
- 02_bonus_efficient-multihead-attention implements and compares different implementation variants of multihead-attention
- 03_understanding-buffers explains the idea behind PyTorch buffers, which are used to implement the causal attention mechanism in chapter 3