Commit
·
df6667b
1
Parent(s):
987aa30
Update README.md
Browse files
README.md
CHANGED
|
@@ -56,24 +56,24 @@ from torch.nn.utils.rnn import pad_sequence
|
|
| 56 |
tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased")
|
| 57 |
|
| 58 |
def tokenize(sample):
|
| 59 |
-
sample["card_name"] = tokenizer(sample["card_name"])["input_ids"]
|
| 60 |
-
sample["type_line"] = tokenizer(sample["type_line"])["input_ids"]
|
| 61 |
-
sample["oracle_text"] = tokenizer(sample["oracle_text"])["input_ids"]
|
| 62 |
-
return sample
|
| 63 |
|
| 64 |
tokenized_dataset = train_dataset.map(tokenize)
|
| 65 |
|
| 66 |
def collate_fn(sequences):
|
| 67 |
-
# Pad the sequences to the maximum length in the batch
|
| 68 |
-
card_names = [torch.tensor(sequence['card_name']) for sequence in sequences]
|
| 69 |
-
type_line = [torch.tensor(sequence['type_line']) for sequence in sequences]
|
| 70 |
-
oracle_text = [torch.tensor(sequence['oracle_text']) for sequence in sequences]
|
| 71 |
|
| 72 |
-
padded_card_name = pad_sequence(card_names, batch_first=True, padding_value=0)
|
| 73 |
-
padded_type_line = pad_sequence(type_line, batch_first=True, padding_value=0)
|
| 74 |
-
padded_oracle_text = pad_sequence(oracle_text, batch_first=True, padding_value=0)
|
| 75 |
|
| 76 |
-
return {'card_name': padded_card_name, 'type_line': padded_type_line, 'padded_oracle_text': padded_oracle_text}
|
| 77 |
|
| 78 |
loader = torch.utils.data.DataLoader(tokenized_dataset, collate_fn=collate_fn, batch_size=4)
|
| 79 |
|
|
|
|
| 56 |
tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased")
|
| 57 |
|
| 58 |
def tokenize(sample):
|
| 59 |
+
sample["card_name"] = tokenizer(sample["card_name"])["input_ids"]
|
| 60 |
+
sample["type_line"] = tokenizer(sample["type_line"])["input_ids"]
|
| 61 |
+
sample["oracle_text"] = tokenizer(sample["oracle_text"])["input_ids"]
|
| 62 |
+
return sample
|
| 63 |
|
| 64 |
tokenized_dataset = train_dataset.map(tokenize)
|
| 65 |
|
| 66 |
def collate_fn(sequences):
|
| 67 |
+
# Pad the sequences to the maximum length in the batch
|
| 68 |
+
card_names = [torch.tensor(sequence['card_name']) for sequence in sequences]
|
| 69 |
+
type_line = [torch.tensor(sequence['type_line']) for sequence in sequences]
|
| 70 |
+
oracle_text = [torch.tensor(sequence['oracle_text']) for sequence in sequences]
|
| 71 |
|
| 72 |
+
padded_card_name = pad_sequence(card_names, batch_first=True, padding_value=0)
|
| 73 |
+
padded_type_line = pad_sequence(type_line, batch_first=True, padding_value=0)
|
| 74 |
+
padded_oracle_text = pad_sequence(oracle_text, batch_first=True, padding_value=0)
|
| 75 |
|
| 76 |
+
return {'card_name': padded_card_name, 'type_line': padded_type_line, 'padded_oracle_text': padded_oracle_text}
|
| 77 |
|
| 78 |
loader = torch.utils.data.DataLoader(tokenized_dataset, collate_fn=collate_fn, batch_size=4)
|
| 79 |
|