mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2026-04-10 12:33:42 +00:00
remove redundant context_length in GQA
This commit is contained in:
@@ -233,7 +233,7 @@
|
||||
"source": [
|
||||
"class GroupedQueryAttention(nn.Module):\n",
|
||||
" def __init__(\n",
|
||||
" self, d_in, d_out, context_length, num_heads,\n",
|
||||
" self, d_in, d_out, num_heads,\n",
|
||||
" num_kv_groups,\n",
|
||||
" dtype=None\n",
|
||||
" ):\n",
|
||||
@@ -320,7 +320,6 @@
|
||||
" self.att = GroupedQueryAttention(\n",
|
||||
" d_in=cfg[\"emb_dim\"],\n",
|
||||
" d_out=cfg[\"emb_dim\"],\n",
|
||||
" context_length=cfg[\"context_length\"],\n",
|
||||
" num_heads=cfg[\"n_heads\"],\n",
|
||||
" num_kv_groups=cfg[\"n_kv_groups\"],\n",
|
||||
" dtype=cfg[\"dtype\"]\n",
|
||||
|
||||
Reference in New Issue
Block a user