Use figure numbers in ch05-7 (#881)

This commit is contained in:
Sebastian Raschka
2025-10-13 16:26:35 -05:00
committed by GitHub
parent bf039ff3dc
commit b969b3ef7a
3 changed files with 54 additions and 55 deletions

View File

@@ -75,7 +75,7 @@
"id": "efd27fcc-2886-47cb-b544-046c2c31f02a",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/chapter-overview.webp\" width=500px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/01.webp\" width=500px>"
]
},
{
@@ -91,7 +91,7 @@
"id": "f67711d4-8391-4fee-aeef-07ea53dd5841",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/mental-model--0.webp\" width=400px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/02.webp\" width=400px>"
]
},
{
@@ -195,7 +195,7 @@
"id": "741881f3-cee0-49ad-b11d-b9df3b3ac234",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/gpt-process.webp\" width=500px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/03.webp\" width=500px>"
]
},
{
@@ -346,7 +346,7 @@
"id": "384d86a9-0013-476c-bb6b-274fd5f20b29",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/proba-to-text.webp\" width=500px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/04.webp\" width=500px>"
]
},
{
@@ -440,7 +440,7 @@
"id": "ad90592f-0d5d-4ec8-9ff5-e7675beab10e",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/proba-index.webp\" width=500px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/06.webp\" width=500px>"
]
},
{
@@ -601,7 +601,7 @@
"id": "5bd24b7f-b760-47ad-bc84-86d13794aa54",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/cross-entropy.webp?123\" width=400px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/07.webp\" width=400px>"
]
},
{
@@ -945,7 +945,7 @@
"id": "46bdaa07-ba96-4ac1-9d71-b3cc153910d9",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/batching.webp\" width=500px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/09.webp\" width=500px>"
]
},
{
@@ -1210,7 +1210,7 @@
"id": "43875e95-190f-4b17-8f9a-35034ba649ec",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/mental-model-1.webp\" width=400px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/10.webp\" width=400px>"
]
},
{
@@ -1231,7 +1231,7 @@
"- In this section, we finally implement the code for training the LLM\n",
"- We focus on a simple training function (if you are interested in augmenting this training function with more advanced techniques, such as learning rate warmup, cosine annealing, and gradient clipping, please refer to [Appendix D](../../appendix-D/01_main-chapter-code))\n",
"\n",
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/train-steps.webp\" width=300px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/11.webp\" width=300px>"
]
},
{
@@ -1464,7 +1464,7 @@
"id": "eb380c42-b31c-4ee1-b8b9-244094537272",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/mental-model-2.webp\" width=350px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/13.webp\" width=350px>"
]
},
{
@@ -1849,7 +1849,7 @@
"id": "7ae6fffd-2730-4abe-a2d3-781fc4836f17",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/topk.webp\" width=500px>\n",
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/15.webp\" width=500px>\n",
"\n",
"- (Please note that the numbers in this figure are truncated to two\n",
"digits after the decimal point to reduce visual clutter. The values in the Softmax row should add up to 1.0.)"
@@ -2060,7 +2060,7 @@
"source": [
"- Training LLMs is computationally expensive, so it's crucial to be able to save and load LLM weights\n",
"\n",
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/mental-model-3.webp\" width=400px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/16.webp\" width=400px>"
]
},
{
@@ -2393,7 +2393,7 @@
"id": "20f19d32-5aae-4176-9f86-f391672c8f0d",
"metadata": {},
"source": [
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/gpt-sizes.webp?timestamp=123\" width=500px>"
"<img src=\"https://sebastianraschka.com/images/LLMs-from-scratch-images/ch05_compressed/17.webp\" width=500px>"
]
},
{
@@ -2627,7 +2627,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.10.16"
"version": "3.13.5"
}
},
"nbformat": 4,