nroggendorff commited on
Commit
42c3bab
·
verified ·
1 Parent(s): cfde5c7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +9 -8
app.py CHANGED
@@ -1,15 +1,16 @@
1
- from transformers import TrainingArguments, AutoConfig, AutoTokenizer, AutoModelForCausalLM
 
2
  import numpy as np
3
- from transformers import LlamaConfig, LlamaForCausalLM
4
- import trl
5
  import torch
 
 
6
  from datasets import load_dataset
7
- from transformers import PreTrainedTokenizerFast
8
- import requests as rq
9
- import gc
10
  from tokenizers import ByteLevelBPETokenizer
 
11
 
12
- dataset = load_dataset("nroggendorff/openhermes", split="train").select(range(int(4e+5)))
13
 
14
  def get_training_corpus():
15
  for i in range(0, len(dataset), 1000):
@@ -98,7 +99,7 @@ print(dataset['text'][2])
98
  args = TrainingArguments(
99
  output_dir="mayo",
100
  num_train_epochs=2,
101
- gradient_accumulation_steps=8,
102
  per_device_train_batch_size=32,
103
  learning_rate=1e-5,
104
  save_steps=100000,
 
1
+ import gc
2
+
3
  import numpy as np
4
+ import requests as rq
 
5
  import torch
6
+
7
+ from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM, PreTrainedTokenizerFast, TrainingArguments
8
  from datasets import load_dataset
9
+
 
 
10
  from tokenizers import ByteLevelBPETokenizer
11
+ import trl
12
 
13
+ dataset = load_dataset("nroggendorff/openhermes", split="train")#.select(range(int(4e+5)))
14
 
15
  def get_training_corpus():
16
  for i in range(0, len(dataset), 1000):
 
99
  args = TrainingArguments(
100
  output_dir="mayo",
101
  num_train_epochs=2,
102
+ gradient_accumulation_steps=16,
103
  per_device_train_batch_size=32,
104
  learning_rate=1e-5,
105
  save_steps=100000,