LLMs-from-scratch/ch05/05_bonus_hparam_tuning
Daniel Kleine dcbdc1d2e5
fixes for code (#206)
* updated .gitignore

* removed unused GELU import

* fixed model_configs, fixed all tensors on same device

* removed unused tiktoken

* update

* update hparam search

* remove redundant tokenizer argument

---------

Co-authored-by: rasbt <mail@sebastianraschka.com>
2024-06-11 20:59:48 -05:00
..
2024-06-11 20:59:48 -05:00
2024-06-11 20:59:48 -05:00
2024-04-10 22:09:46 -04:00

Optimizing Hyperparameters for Pretraining

The hparam_search.py script, based on the extended training function in Appendix D: Adding Bells and Whistles to the Training Loop, is designed to find optimal hyperparameters via grid search.

Note

This script will take a long time to run. You may want to reduce the number of hyperparameter configurations explored in the HPARAM_GRID dictionary at the top.