mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2025-11-02 02:41:00 +00:00
* typo fix experiment 5 is the same size as experiment 1, both are 124 Million. and experiment 6 is 355M ~3x 124M. * typo fix all layer FT vs all layer FT, "slightly worse by 1.3% " indicates it's exp 5 vs exp 9 * exp 5 vs 9 * Update ch06/02_bonus_additional-experiments/README.md --------- Co-authored-by: Sebastian Raschka <mail@sebastianraschka.com>
Chapter 6: Finetuning for Classification
Main Chapter Code
- 01_main-chapter-code contains the main chapter code
Bonus Materials
- 02_bonus_additional-experiments includes additional experiments (e.g., training the last vs first token, extending the input length, etc.)
- 03_bonus_imdb-classification compares the LLM from chapter 6 with other models on a 50k IMDB movie review sentiment classification dataset