Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,15 +1,16 @@
|
|
| 1 |
-
|
|
|
|
| 2 |
import numpy as np
|
| 3 |
-
|
| 4 |
-
import trl
|
| 5 |
import torch
|
|
|
|
|
|
|
| 6 |
from datasets import load_dataset
|
| 7 |
-
|
| 8 |
-
import requests as rq
|
| 9 |
-
import gc
|
| 10 |
from tokenizers import ByteLevelBPETokenizer
|
|
|
|
| 11 |
|
| 12 |
-
dataset = load_dataset("nroggendorff/openhermes", split="train").select(range(int(4e+5)))
|
| 13 |
|
| 14 |
def get_training_corpus():
|
| 15 |
for i in range(0, len(dataset), 1000):
|
|
@@ -98,7 +99,7 @@ print(dataset['text'][2])
|
|
| 98 |
args = TrainingArguments(
|
| 99 |
output_dir="mayo",
|
| 100 |
num_train_epochs=2,
|
| 101 |
-
gradient_accumulation_steps=
|
| 102 |
per_device_train_batch_size=32,
|
| 103 |
learning_rate=1e-5,
|
| 104 |
save_steps=100000,
|
|
|
|
| 1 |
+
import gc
|
| 2 |
+
|
| 3 |
import numpy as np
|
| 4 |
+
import requests as rq
|
|
|
|
| 5 |
import torch
|
| 6 |
+
|
| 7 |
+
from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM, PreTrainedTokenizerFast, TrainingArguments
|
| 8 |
from datasets import load_dataset
|
| 9 |
+
|
|
|
|
|
|
|
| 10 |
from tokenizers import ByteLevelBPETokenizer
|
| 11 |
+
import trl
|
| 12 |
|
| 13 |
+
dataset = load_dataset("nroggendorff/openhermes", split="train")#.select(range(int(4e+5)))
|
| 14 |
|
| 15 |
def get_training_corpus():
|
| 16 |
for i in range(0, len(dataset), 1000):
|
|
|
|
| 99 |
args = TrainingArguments(
|
| 100 |
output_dir="mayo",
|
| 101 |
num_train_epochs=2,
|
| 102 |
+
gradient_accumulation_steps=16,
|
| 103 |
per_device_train_batch_size=32,
|
| 104 |
learning_rate=1e-5,
|
| 105 |
save_steps=100000,
|