msinghy commited on
Commit
0b59fdf
·
verified ·
1 Parent(s): 4167a41

Training in progress, epoch 5, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf7f1711371cdf22625cfd71373d0e13667f9262db729700ebba8bff281e302f
3
  size 6403448
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e3dbdf3aa088737ce1cdad6ea130252698559d75de369f4db5b9d8dac0347ff
3
  size 6403448
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa22af2fc608eb3be5b6d7756926317b864996be333d25aedbb9e48d2d5bd10f
3
  size 12867066
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf90bfc28626b9b635a21b6c7af37c270a72edf31a353533fd4ccef7270c5d23
3
  size 12867066
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a0a24216fd4977ce1d76fe035f83db58fced950e69f3285260aec3babc9a654
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10e3856931a60333566c32feed9f656e65195d769bc3d7827e3d40c2fac2012a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:492f28b6539849c91c4a713593f718b0248ba49625b6a14592a3de84ad5d56c7
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d99598dae57d4b1975eeb0f5c3edb922a43a02c02fef0ccb682d848ccbb64cfa
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 52,
3
- "best_metric": 5.0411696434021,
4
- "best_model_checkpoint": "My-Nietzsche-Model/checkpoint-52",
5
- "epoch": 4.0,
6
  "eval_steps": 500,
7
- "global_step": 52,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -40,6 +40,14 @@
40
  "eval_samples_per_second": 3.212,
41
  "eval_steps_per_second": 0.803,
42
  "step": 52
 
 
 
 
 
 
 
 
43
  }
44
  ],
45
  "logging_steps": 500,
@@ -54,12 +62,12 @@
54
  "should_evaluate": false,
55
  "should_log": false,
56
  "should_save": true,
57
- "should_training_stop": false
58
  },
59
  "attributes": {}
60
  }
61
  },
62
- "total_flos": 2494553191833600.0,
63
  "train_batch_size": 4,
64
  "trial_name": null,
65
  "trial_params": null
 
1
  {
2
+ "best_global_step": 65,
3
+ "best_metric": 5.032853603363037,
4
+ "best_model_checkpoint": "My-Nietzsche-Model/checkpoint-65",
5
+ "epoch": 5.0,
6
  "eval_steps": 500,
7
+ "global_step": 65,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
40
  "eval_samples_per_second": 3.212,
41
  "eval_steps_per_second": 0.803,
42
  "step": 52
43
+ },
44
+ {
45
+ "epoch": 5.0,
46
+ "eval_loss": 5.032853603363037,
47
+ "eval_runtime": 6.187,
48
+ "eval_samples_per_second": 3.233,
49
+ "eval_steps_per_second": 0.808,
50
+ "step": 65
51
  }
52
  ],
53
  "logging_steps": 500,
 
62
  "should_evaluate": false,
63
  "should_log": false,
64
  "should_save": true,
65
+ "should_training_stop": true
66
  },
67
  "attributes": {}
68
  }
69
  },
70
+ "total_flos": 3118191489792000.0,
71
  "train_batch_size": 4,
72
  "trial_name": null,
73
  "trial_params": null