Skip to content

Commit 81eed9a

Browse files
d-kleinerasbt
andauthored
updated RoPE statement (#423)
* updated RoPE statement * updated .gitignore * Update ch05/07_gpt_to_llama/converting-gpt-to-llama2.ipynb --------- Co-authored-by: Sebastian Raschka <[email protected]>
1 parent b5f2aa3 commit 81eed9a

File tree

2 files changed

+3
-1
lines changed

2 files changed

+3
-1
lines changed

.gitignore

+2
Original file line numberDiff line numberDiff line change
@@ -44,6 +44,8 @@ ch05/07_gpt_to_llama/Llama-3.1-8B
4444
ch05/07_gpt_to_llama/Llama-3.1-8B-Instruct
4545
ch05/07_gpt_to_llama/Llama-3.2-1B
4646
ch05/07_gpt_to_llama/Llama-3.2-1B-Instruct
47+
ch05/07_gpt_to_llama/Llama-3.2-3B
48+
ch05/07_gpt_to_llama/Llama-3.2-3B-Instruct
4749

4850
ch06/01_main-chapter-code/gpt2
4951
ch06/02_bonus_additional-experiments/gpt2

ch05/07_gpt_to_llama/converting-gpt-to-llama2.ipynb

+1-1
Original file line numberDiff line numberDiff line change
@@ -409,7 +409,7 @@
409409
"self.pos_emb = nn.Embedding(cfg[\"context_length\"], cfg[\"emb_dim\"])\n",
410410
"```\n",
411411
"\n",
412-
"- Instead of these absolute positional embeddings, Llama uses relative positional embeddings, called rotary position embeddings (RoPE for short)\n",
412+
"- Unlike traditional absolute positional embeddings, Llama uses rotary position embeddings (RoPE), which enable it to capture both absolute and relative positional information simultaneously\n",
413413
"- The reference paper for RoPE is [RoFormer: Enhanced Transformer with Rotary Position Embedding (2021)](https://arxiv.org/abs/2104.09864)"
414414
]
415415
},

0 commit comments

Comments
 (0)