LLMs-from-scratch/ch03/02_bonus_efficient-multihead-attention
Sebastian Raschka 0528446584 Make code more consistent and add projection layer (#131)
* Make code more consistent and add projection

* remove redundant buffer
2024-04-26 17:13:08 -05:00
..
2024-04-04 07:58:41 -05:00
2024-03-06 08:30:32 -06:00

More Efficient Multi-Head Attention Implementations