mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2025-08-29 02:50:15 +00:00

* Auto download DPO dataset if not already available in path * update tests to account for latest HF transformers release in unit tests * pep 8
Chapter 7: Finetuning to Follow Instructions
-
create-preference-data-ollama.ipynb: A notebook that creates a synthetic dataset for preference finetuning dataset using Llama 3.1 and Ollama
-
dpo-from-scratch.ipynb: This notebook implements Direct Preference Optimization (DPO) for LLM alignment