mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2025-09-01 20:38:11 +00:00
8 lines
366 B
Markdown
8 lines
366 B
Markdown
# Chapter 7: Finetuning to Follow Instructions
|
|
|
|
- [create-preference-data-ollama.ipynb](create-preference-data-ollama.ipynb): A notebook that creates a synthetic dataset for preference finetuning dataset using Llama 3.1 and Ollama
|
|
|
|
- [dpo-from-scratch.ipynb](dpo-from-scratch.ipynb): This notebook implements Direct Preference Optimization (DPO) for LLM alignment
|
|
|
|
|