Skip to content

Commit 15d2f35

Browse files
committed
no rotary embedding for transformers within CLIP model
1 parent 3df569d commit 15d2f35

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

dalle_pytorch/dalle_pytorch.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -247,7 +247,7 @@ def __init__(
247247
super().__init__()
248248
self.text_emb = nn.Embedding(num_text_tokens, dim_text)
249249
self.text_pos_emb = nn.Embedding(text_seq_len, dim_text)
250-
self.text_transformer = Transformer(causal = False, seq_len = text_seq_len, dim = dim_text, depth = text_enc_depth, heads = text_heads)
250+
self.text_transformer = Transformer(causal = False, seq_len = text_seq_len, dim = dim_text, depth = text_enc_depth, heads = text_heads, rotary_emb = False)
251251
self.to_text_latent = nn.Linear(dim_text, dim_latent, bias = False)
252252

253253
assert visual_image_size % visual_patch_size == 0, 'Image dimensions must be divisible by the patch size.'
@@ -257,7 +257,7 @@ def __init__(
257257
self.visual_patch_size = visual_patch_size
258258
self.to_visual_embedding = nn.Linear(patch_dim, dim_image)
259259
self.visual_pos_emb = nn.Embedding(num_patches, dim_image)
260-
self.visual_transformer = Transformer(causal = False, seq_len = num_patches, dim = dim_image, depth = visual_enc_depth, heads = visual_heads)
260+
self.visual_transformer = Transformer(causal = False, seq_len = num_patches, dim = dim_image, depth = visual_enc_depth, heads = visual_heads, rotary_emb = False)
261261
self.to_visual_latent = nn.Linear(dim_image, dim_latent, bias = False)
262262

263263
self.temperature = nn.Parameter(torch.tensor(1.))

setup.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
name = 'dalle-pytorch',
55
packages = find_packages(),
66
include_package_data = True,
7-
version = '1.0.7',
7+
version = '1.0.8',
88
license='MIT',
99
description = 'DALL-E - Pytorch',
1010
author = 'Phil Wang',

0 commit comments

Comments
 (0)