mirror of
				https://github.com/rasbt/LLMs-from-scratch.git
				synced 2025-11-03 19:30:26 +00:00 
			
		
		
		
	
		
			
				
	
	
	
		
			938 B
		
	
	
	
	
	
	
	
			
		
		
	
	
			938 B
		
	
	
	
	
	
	
	
More Efficient Multi-Head Attention Implementations
- mha-implementations.ipynb contains and compares different implementations of multi-head attention
 
Summary
The figures below summarize the performance benchmarks (lower is better).
Forward pass only
Forward and backward pass


