21 Commits

Author SHA1 Message Date
Sebastian Raschka
be5e2a3331 Readability and code quality improvements (#959)
* Consistent dataset naming

* consistent section headers
2026-02-17 18:44:56 -06:00
casinca
9c4be478f8 Optional weight tying for Qwen3 and Llama3.2 pretraining (#949)
* optional weight tying for Qwen3 and Llama3.2

* typo
2026-01-14 09:07:04 -06:00
Sebastian Raschka
90e0f3cc15 Chapter 5 with alternative LLMs (Qwen3, Llama 3) (#943)
* ch05-alt-llms

* fix double import

* style fix

* Fix urls

* Fix urls

* Fix urls
2026-01-09 14:58:20 -06:00
Sebastian Raschka
9bc827ea7e Numerically stable generate on mps (#849)
* Numerically stable generate on mps

* add file
2025-09-26 22:42:44 -05:00
Sebastian Raschka
b6cd0a312f More efficient angles computation in RoPE (#830) 2025-09-16 03:23:33 +00:00
Sebastian Raschka
8add26cbe9 Improve weight tying handling (#826)
* Improve weight tying handling

* fix
2025-09-14 15:46:48 -05:00
Daniel Kleine
c2cfb47b1a fixed gqa qkv code comments (#660) 2025-06-13 08:21:28 -05:00
Sebastian Raschka
c4cde1c21b Reduce Llama 3 RoPE memory requirements (#658)
* Llama3 from scratch improvements

* Fix Llama 3 expensive RoPE memory issue

* updates

* update package

* benchmark

* remove unused rescale_theta
2025-06-12 11:08:02 -05:00
Sebastian Raschka
67e0680210 Disable mask saving as weight in Llama 3 model (#604)
* Disable mask saving as weight

* update pixi

* update pixi
2025-04-06 09:33:36 -05:00
Sebastian Raschka
d4c8d8f2c9 Fix Llama language typo in bonus materials (#597) 2025-04-02 21:41:36 -05:00
Sebastian Raschka
a08d7aaa84 Uv workflow improvements (#531)
* Uv workflow improvements

* Uv workflow improvements

* linter improvements

* pytproject.toml fixes

* pytproject.toml fixes

* pytproject.toml fixes

* pytproject.toml fixes

* pytproject.toml fixes

* pytproject.toml fixes

* windows fixes

* windows fixes

* windows fixes

* windows fixes

* windows fixes

* windows fixes

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix

* win32 fix
2025-02-16 13:16:51 -06:00
casinca
bb31de8999 [minor] typo & comments (#441)
* typo & comment

- safe -> save
- commenting code: batch_size, seq_len = in_idx.shape

* comment

- adding # NEW for assert num_heads % num_kv_groups == 0

* update memory wording

---------

Co-authored-by: rasbt <mail@sebastianraschka.com>
2024-11-18 19:52:42 +09:00
Daniel Kleine
e8c2f962e9 minor fixes: Llama 3.2 standalone (#420)
* minor fixes

* reformat rope base as float

---------

Co-authored-by: rasbt <mail@sebastianraschka.com>
2024-10-25 21:08:06 -05:00
Sebastian Raschka
1516de54a5 RoPE theta rescaling (#419)
* rope fixes

* update

* update

* cleanup
2024-10-25 15:27:23 -05:00
Daniel Kleine
d38083c401 Updated Llama 2 to 3 paths (#413)
* llama 2 and 3 path fixes

* updated llama 3, 3.1 and 3.2 paths

* updated .gitignore

* Typo fix

---------

Co-authored-by: Sebastian Raschka <mail@sebastianraschka.com>
2024-10-24 07:40:08 -05:00
Sebastian Raschka
7cd6a670ed RoPE updates (#412)
* RoPE updates

* Apply suggestions from code review

* updates

* updates

* updates
2024-10-23 18:07:49 -05:00
Sebastian Raschka
534a704364 RoPE increase (#407) 2024-10-21 19:58:38 -05:00
Sebastian Raschka
1eb0b3810a Introduce buffers to improve Llama 3.2 efficiency (#389)
* Introduce buffers to improve Llama 3.2 efficiency

* update

* update
2024-10-06 12:49:04 -05:00
Daniel Kleine
a0c0c765a8 fixed Llama 2 to 3.2 NBs (#388)
* updated requirements

* fixes llama2 to llama3

* fixed llama 3.2 standalone

* fixed typo

* fixed rope formula

* Update requirements-extra.txt

* Update ch05/07_gpt_to_llama/converting-llama2-to-llama3.ipynb

* Update ch05/07_gpt_to_llama/converting-llama2-to-llama3.ipynb

* Update ch05/07_gpt_to_llama/standalone-llama32.ipynb

---------

Co-authored-by: Sebastian Raschka <mail@sebastianraschka.com>
2024-10-06 09:56:55 -05:00
Sebastian Raschka
0972ded530 Add a note about weight tying in Llama 3.2 (#386) 2024-10-05 09:20:54 -05:00
Sebastian Raschka
b44096acef Implement Llama 3.2 (#383) 2024-10-05 07:30:47 -05:00