Skip to content

Commit 152a087

Browse files
casincarasbt
andauthored
removing unused RoPE parameters (#590)
* removing unused RoPE parameters * remove redundant context_length in GQA --------- Co-authored-by: Sebastian Raschka <[email protected]>
1 parent 2228037 commit 152a087

File tree

1 file changed

+1
-6
lines changed

1 file changed

+1
-6
lines changed

ch05/07_gpt_to_llama/standalone-llama32-mem-opt.ipynb

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -233,10 +233,8 @@
233233
"source": [
234234
"class GroupedQueryAttention(nn.Module):\n",
235235
" def __init__(\n",
236-
" self, d_in, d_out, context_length, num_heads,\n",
236+
" self, d_in, d_out, num_heads,\n",
237237
" num_kv_groups,\n",
238-
" rope_base=10_000,\n",
239-
" rope_config=None,\n",
240238
" dtype=None\n",
241239
" ):\n",
242240
" super().__init__()\n",
@@ -322,11 +320,8 @@
322320
" self.att = GroupedQueryAttention(\n",
323321
" d_in=cfg[\"emb_dim\"],\n",
324322
" d_out=cfg[\"emb_dim\"],\n",
325-
" context_length=cfg[\"context_length\"],\n",
326323
" num_heads=cfg[\"n_heads\"],\n",
327324
" num_kv_groups=cfg[\"n_kv_groups\"],\n",
328-
" rope_base=cfg[\"rope_base\"],\n",
329-
" rope_config=cfg[\"rope_freq\"],\n",
330325
" dtype=cfg[\"dtype\"]\n",
331326
" )\n",
332327
" self.ff = FeedForward(cfg)\n",

0 commit comments

Comments
 (0)