LLMs-from-scratch/ch05/01_main-chapter-code
James Holcombe 05718c6b94
Use instance tokenizer (#116)
* Use instance tokenizer

* consistency updates

---------

Co-authored-by: Sebastian Raschka <mail@sebastianraschka.com>
2024-04-10 21:16:19 -04:00
..
2024-04-07 06:03:41 -05:00
2024-04-05 07:24:46 -05:00
2024-04-05 07:24:46 -05:00

Chapter 5: Pretraining on Unlabeled Data

  • ch05.ipynb contains all the code as it appears in the chapter
  • previous_chapters.py is a Python module that contains the MultiHeadAttention module from the previous chapter, which we import in ch05.ipynb to pretrain the GPT model
  • train.py is a standalone Python script file with the code that we implemented in ch05.ipynb to train the GPT model
  • generate.py is a standalone Python script file with the code that we implemented in ch05.ipynb to load and use the pretrained model weights from OpenAI