8 lines
366 B
Markdown
Raw Normal View History

2024-06-08 10:38:41 -05:00
# Chapter 7: Finetuning to Follow Instructions
2024-06-05 08:21:28 -05:00
- [create-preference-data-ollama.ipynb](create-preference-data-ollama.ipynb): A notebook that creates a synthetic dataset for preference finetuning dataset using Llama 3.1 and Ollama
2024-06-23 08:25:01 -05:00
- [dpo-from-scratch.ipynb](dpo-from-scratch.ipynb): This notebook implements Direct Preference Optimization (DPO) for LLM alignment