From a10d0ea661804be1ecdd32ea4a72a67bcd68a1bd Mon Sep 17 00:00:00 2001 From: casinca <47400729+casinca@users.noreply.github.com> Date: Mon, 31 Mar 2025 19:19:27 +0200 Subject: [PATCH] removing unused RoPE parameters --- ch05/07_gpt_to_llama/standalone-llama32-mem-opt.ipynb | 4 ---- 1 file changed, 4 deletions(-) diff --git a/ch05/07_gpt_to_llama/standalone-llama32-mem-opt.ipynb b/ch05/07_gpt_to_llama/standalone-llama32-mem-opt.ipynb index 02b61e5..284c0ec 100644 --- a/ch05/07_gpt_to_llama/standalone-llama32-mem-opt.ipynb +++ b/ch05/07_gpt_to_llama/standalone-llama32-mem-opt.ipynb @@ -235,8 +235,6 @@ " def __init__(\n", " self, d_in, d_out, context_length, num_heads,\n", " num_kv_groups,\n", - " rope_base=10_000,\n", - " rope_config=None,\n", " dtype=None\n", " ):\n", " super().__init__()\n", @@ -325,8 +323,6 @@ " context_length=cfg[\"context_length\"],\n", " num_heads=cfg[\"n_heads\"],\n", " num_kv_groups=cfg[\"n_kv_groups\"],\n", - " rope_base=cfg[\"rope_base\"],\n", - " rope_config=cfg[\"rope_freq\"],\n", " dtype=cfg[\"dtype\"]\n", " )\n", " self.ff = FeedForward(cfg)\n",