gokuls commited on
Commit
06c208d
·
1 Parent(s): 67df07a

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ trainer_state.json filter=lfs diff=lfs merge=lfs -text
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.98,
3
+ "eval_accuracy": 0.5626417305418694,
4
+ "eval_loss": 2.3996434211730957,
5
+ "eval_runtime": 2690.6613,
6
+ "eval_samples": 308356,
7
+ "eval_samples_per_second": 114.602,
8
+ "eval_steps_per_second": 3.582,
9
+ "perplexity": 11.019246450045694,
10
+ "train_loss": 2.6296509271005224,
11
+ "train_runtime": 187194.7694,
12
+ "train_samples": 5858758,
13
+ "train_samples_per_second": 156.488,
14
+ "train_steps_per_second": 4.89
15
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.98,
3
+ "eval_accuracy": 0.5626417305418694,
4
+ "eval_loss": 2.3996434211730957,
5
+ "eval_runtime": 2690.6613,
6
+ "eval_samples": 308356,
7
+ "eval_samples_per_second": 114.602,
8
+ "eval_steps_per_second": 3.582,
9
+ "perplexity": 11.019246450045694
10
+ }
logs/events.out.tfevents.1689863915.serv-9207.462220.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f83750ae1af0f05e7d9c3f17c08dde63b73e90a14279fc6bb0e662e5f0b3960e
3
+ size 369
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.98,
3
+ "train_loss": 2.6296509271005224,
4
+ "train_runtime": 187194.7694,
5
+ "train_samples": 5858758,
6
+ "train_samples_per_second": 156.488,
7
+ "train_steps_per_second": 4.89
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f22c42a1a452870793294a70ae99ce5bacc8e22922d24db447e6ef44b68e614e
3
+ size 21868720