mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2026-04-10 12:33:42 +00:00
distinguish better between main chapter code and bonus materials
This commit is contained in:
@@ -1,6 +1,11 @@
|
||||
# Chapter 5: Pretraining on Unlabeled Data
|
||||
|
||||
## Main Chapter Code
|
||||
|
||||
- [01_main-chapter-code](01_main-chapter-code) contains the main chapter code
|
||||
|
||||
## Bonus Materials
|
||||
|
||||
- [02_alternative_weight_loading](02_alternative_weight_loading) contains code to load the GPT model weights from alternative places in case the model weights become unavailable from OpenAI
|
||||
- [03_bonus_pretraining_on_gutenberg](03_bonus_pretraining_on_gutenberg) contains code to pretrain the LLM longer on the whole corpus of books from Project Gutenberg
|
||||
- [04_learning_rate_schedulers](04_learning_rate_schedulers) contains code implementing a more sophisticated training function including learning rate schedulers and gradient clipping
|
||||
|
||||
Reference in New Issue
Block a user