l;l
Browse files- modelLM.py +3 -3
modelLM.py
CHANGED
@@ -33,13 +33,13 @@ class OBILanguageModel(PreTrainedModel):
|
|
33 |
tok_emb = self.token_embedding_table(idx)
|
34 |
pos_emb = None # Initialize pos_emb to None
|
35 |
try:
|
36 |
-
pos_emb = self.position_embedding_table(torch.arange(
|
37 |
except IndexError as e:
|
38 |
# Handle the IndexError by initializing pos_emb with zeros
|
39 |
print(f"IndexError: {e}")
|
40 |
-
print(f"idx.size(
|
41 |
print(f"Positional embedding table shape: {self.position_embedding_table.weight.shape}")
|
42 |
-
pos_emb = torch.zeros((
|
43 |
|
44 |
x = tok_emb + pos_emb
|
45 |
x = self.transformer(x, x)
|
|
|
33 |
tok_emb = self.token_embedding_table(idx)
|
34 |
pos_emb = None # Initialize pos_emb to None
|
35 |
try:
|
36 |
+
pos_emb = self.position_embedding_table(torch.arange(idx.size(1), device='cpu'))
|
37 |
except IndexError as e:
|
38 |
# Handle the IndexError by initializing pos_emb with zeros
|
39 |
print(f"IndexError: {e}")
|
40 |
+
print(f"idx.size(1): {idx.size(1)}")
|
41 |
print(f"Positional embedding table shape: {self.position_embedding_table.weight.shape}")
|
42 |
+
pos_emb = torch.zeros((idx.size(1), self.config.hidden_size), device=device)
|
43 |
|
44 |
x = tok_emb + pos_emb
|
45 |
x = self.transformer(x, x)
|