Spaces:
Paused
Paused
Update train.py
Browse files
train.py
CHANGED
|
@@ -9,11 +9,11 @@ from tokenizers import ByteLevelBPETokenizer
|
|
| 9 |
|
| 10 |
MAX_SEQ_LENGTH = 512
|
| 11 |
BATCH_SIZE = 128
|
| 12 |
-
EPOCHS =
|
| 13 |
LEARNING_RATE = 2e-2
|
| 14 |
FACTOR = 256
|
| 15 |
VOCAB_SIZE = 32000
|
| 16 |
-
INPUT_DATASET = "nroggendorff/
|
| 17 |
OUTPUT_REPO = "smallama"
|
| 18 |
FP16 = True
|
| 19 |
WARMUP_STEPS = 20
|
|
@@ -23,7 +23,7 @@ CLIPPING = 1.0
|
|
| 23 |
PUSH_TO_HUB = True
|
| 24 |
|
| 25 |
def load_data():
|
| 26 |
-
dataset = load_dataset(INPUT_DATASET, split="train").select(range(int(5e+
|
| 27 |
return dataset
|
| 28 |
|
| 29 |
def create_tokenizer(training_corpus):
|
|
|
|
| 9 |
|
| 10 |
MAX_SEQ_LENGTH = 512
|
| 11 |
BATCH_SIZE = 128
|
| 12 |
+
EPOCHS = 2
|
| 13 |
LEARNING_RATE = 2e-2
|
| 14 |
FACTOR = 256
|
| 15 |
VOCAB_SIZE = 32000
|
| 16 |
+
INPUT_DATASET = "nroggendorff/elephant"
|
| 17 |
OUTPUT_REPO = "smallama"
|
| 18 |
FP16 = True
|
| 19 |
WARMUP_STEPS = 20
|
|
|
|
| 23 |
PUSH_TO_HUB = True
|
| 24 |
|
| 25 |
def load_data():
|
| 26 |
+
dataset = load_dataset(INPUT_DATASET, split="train").select(range(int(2.5e+6)))
|
| 27 |
return dataset
|
| 28 |
|
| 29 |
def create_tokenizer(training_corpus):
|