mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2025-08-08 16:52:43 +00:00

* fixed issues, updated .gitignore * added closing paren * fixed CEL spelling * fixed more minor issues * Update ch07/01_main-chapter-code/ch07.ipynb * Update ch07/04_preference-tuning-with-dpo/dpo-from-scratch.ipynb * Update ch07/04_preference-tuning-with-dpo/dpo-from-scratch.ipynb * Update ch07/04_preference-tuning-with-dpo/dpo-from-scratch.ipynb --------- Co-authored-by: Sebastian Raschka <mail@sebastianraschka.com>
Chapter 7: Finetuning to Follow Instructions
Main Chapter Code
- 01_main-chapter-code contains the main chapter code and exercise solutions
Bonus Materials
-
02_dataset-utilities contains utility code that can be used for preparing an instruction dataset.
-
03_model-evaluation contains utility code for evaluating instruction responses using a local Llama 3 model and the GPT-4 API.
-
04_preference-tuning-with-dpo implements code for preference finetuning with Direct Preference Optimization (DPO)
-
05_dataset-generation contains code to generate synthetic datasets for instruction finetuning