From 3654571184b69ac79963092f24b115cdb8a14eeb Mon Sep 17 00:00:00 2001 From: Sebastian Raschka Date: Sun, 6 Apr 2025 16:46:53 -0500 Subject: [PATCH] align formulas in notes with code (#605) --- ch03/01_main-chapter-code/ch03.ipynb | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/ch03/01_main-chapter-code/ch03.ipynb b/ch03/01_main-chapter-code/ch03.ipynb index 9fc9ffd..148ce26 100644 --- a/ch03/01_main-chapter-code/ch03.ipynb +++ b/ch03/01_main-chapter-code/ch03.ipynb @@ -797,9 +797,9 @@ "- Implementing the self-attention mechanism step by step, we will start by introducing the three training weight matrices $W_q$, $W_k$, and $W_v$\n", "- These three matrices are used to project the embedded input tokens, $x^{(i)}$, into query, key, and value vectors via matrix multiplication:\n", "\n", - " - Query vector: $q^{(i)} = W_q \\,x^{(i)}$\n", - " - Key vector: $k^{(i)} = W_k \\,x^{(i)}$\n", - " - Value vector: $v^{(i)} = W_v \\,x^{(i)}$\n" + " - Query vector: $q^{(i)} = x^{(i)}\\,W_q $\n", + " - Key vector: $k^{(i)} = x^{(i)}\\,W_k $\n", + " - Value vector: $v^{(i)} = x^{(i)}\\,W_v $\n" ] }, {