Skip to content

Commit d37ddb6

Browse files
authored
Fix code comment: embed_dim -> d_out (#698)
1 parent 01be5a4 commit d37ddb6

File tree

3 files changed

+4
-4
lines changed

3 files changed

+4
-4
lines changed

ch03/02_bonus_efficient-multihead-attention/mha-implementations.ipynb

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -352,7 +352,7 @@
352352
" def __init__(self, d_in, d_out, num_heads, context_length, dropout=0.0, qkv_bias=False):\n",
353353
" super().__init__()\n",
354354
"\n",
355-
" assert d_out % num_heads == 0, \"embed_dim is indivisible by num_heads\"\n",
355+
" assert d_out % num_heads == 0, \"d_out is indivisible by num_heads\"\n",
356356
"\n",
357357
" self.num_heads = num_heads\n",
358358
" self.context_length = context_length\n",
@@ -588,7 +588,7 @@
588588
" def __init__(self, d_in, d_out, num_heads, context_length, dropout=0.0, qkv_bias=False):\n",
589589
" super().__init__()\n",
590590
"\n",
591-
" assert d_out % num_heads == 0, \"embed_dim is indivisible by num_heads\"\n",
591+
" assert d_out % num_heads == 0, \"d_out is indivisible by num_heads\"\n",
592592
"\n",
593593
" self.num_heads = num_heads\n",
594594
" self.context_length = context_length\n",

ch05/10_llm-training-speed/01_opt_single_gpu.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ class PyTorchMultiHeadAttention(nn.Module):
6565
def __init__(self, d_in, d_out, num_heads, dropout=0.0, qkv_bias=False):
6666
super().__init__()
6767

68-
assert d_out % num_heads == 0, "embed_dim is indivisible by num_heads"
68+
assert d_out % num_heads == 0, "d_out is indivisible by num_heads"
6969

7070
self.num_heads = num_heads
7171
self.head_dim = d_out // num_heads

ch05/10_llm-training-speed/02_opt_multi_gpu_ddp.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -108,7 +108,7 @@ class PyTorchMultiHeadAttention(nn.Module):
108108
def __init__(self, d_in, d_out, num_heads, dropout=0.0, qkv_bias=False):
109109
super().__init__()
110110

111-
assert d_out % num_heads == 0, "embed_dim is indivisible by num_heads"
111+
assert d_out % num_heads == 0, "d_out is indivisible by num_heads"
112112

113113
self.num_heads = num_heads
114114
self.head_dim = d_out // num_heads

0 commit comments

Comments
 (0)