besimray commited on
Commit
9fa9db3
·
verified ·
1 Parent(s): dda3dc7

Training in progress, step 10, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9880f9a14e7e1b16646467ab0d161d117f1598c30881f29672a2b60ced9327ee
3
  size 67662840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51d535ba0e660420acdf72e7a6a0f5e895765af8e206cc752a5c2a10f730e283
3
  size 67662840
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb444bf5ba451c33cc607315d448f7fa02a9ff9b3bbe6e66bca96ebcbf0fa021
3
  size 34607610
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86335376bbc378ffc6db3cd86237f4e02e04331ccfd8606fc3c0d43db50a728d
3
  size 34607610
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eca66d4a74945f7af156bf782963ee63f197435058aae7c08031504382213d7e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:340f6406b96d7e0f17315436c2bccb605d75c9d254ec958f91a4f659e4ffcdd7
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85e9b4e25668eb135b65c8a2efd824713d5ec7e9a404d37d2fd3021c6e5610a3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:284a8517b2310fb02b51d6b8ddd318d3f6e139475d47fc950976ac1287debd43
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.07352941176470588,
5
  "eval_steps": 1,
6
- "global_step": 5,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -82,6 +82,81 @@
82
  "eval_samples_per_second": 29.814,
83
  "eval_steps_per_second": 4.472,
84
  "step": 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
85
  }
86
  ],
87
  "logging_steps": 1,
@@ -96,12 +171,12 @@
96
  "should_evaluate": false,
97
  "should_log": false,
98
  "should_save": true,
99
- "should_training_stop": false
100
  },
101
  "attributes": {}
102
  }
103
  },
104
- "total_flos": 314029414809600.0,
105
  "train_batch_size": 7,
106
  "trial_name": null,
107
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.14705882352941177,
5
  "eval_steps": 1,
6
+ "global_step": 10,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
82
  "eval_samples_per_second": 29.814,
83
  "eval_steps_per_second": 4.472,
84
  "step": 5
85
+ },
86
+ {
87
+ "epoch": 0.08823529411764706,
88
+ "grad_norm": 1.395002841949463,
89
+ "learning_rate": 3e-06,
90
+ "loss": 1.4327,
91
+ "step": 6
92
+ },
93
+ {
94
+ "epoch": 0.08823529411764706,
95
+ "eval_loss": 1.2674007415771484,
96
+ "eval_runtime": 2.5632,
97
+ "eval_samples_per_second": 39.013,
98
+ "eval_steps_per_second": 5.852,
99
+ "step": 6
100
+ },
101
+ {
102
+ "epoch": 0.10294117647058823,
103
+ "grad_norm": 1.0900819301605225,
104
+ "learning_rate": 3.5000000000000004e-06,
105
+ "loss": 1.2795,
106
+ "step": 7
107
+ },
108
+ {
109
+ "epoch": 0.10294117647058823,
110
+ "eval_loss": 1.269196629524231,
111
+ "eval_runtime": 2.4479,
112
+ "eval_samples_per_second": 40.852,
113
+ "eval_steps_per_second": 6.128,
114
+ "step": 7
115
+ },
116
+ {
117
+ "epoch": 0.11764705882352941,
118
+ "grad_norm": 1.7616393566131592,
119
+ "learning_rate": 4.000000000000001e-06,
120
+ "loss": 1.4695,
121
+ "step": 8
122
+ },
123
+ {
124
+ "epoch": 0.11764705882352941,
125
+ "eval_loss": 1.267441749572754,
126
+ "eval_runtime": 2.4514,
127
+ "eval_samples_per_second": 40.792,
128
+ "eval_steps_per_second": 6.119,
129
+ "step": 8
130
+ },
131
+ {
132
+ "epoch": 0.1323529411764706,
133
+ "grad_norm": 1.3893433809280396,
134
+ "learning_rate": 4.5e-06,
135
+ "loss": 1.4243,
136
+ "step": 9
137
+ },
138
+ {
139
+ "epoch": 0.1323529411764706,
140
+ "eval_loss": 1.2656891345977783,
141
+ "eval_runtime": 3.1314,
142
+ "eval_samples_per_second": 31.934,
143
+ "eval_steps_per_second": 4.79,
144
+ "step": 9
145
+ },
146
+ {
147
+ "epoch": 0.14705882352941177,
148
+ "grad_norm": 1.2425161600112915,
149
+ "learning_rate": 5e-06,
150
+ "loss": 1.4099,
151
+ "step": 10
152
+ },
153
+ {
154
+ "epoch": 0.14705882352941177,
155
+ "eval_loss": 1.2637847661972046,
156
+ "eval_runtime": 2.4954,
157
+ "eval_samples_per_second": 40.074,
158
+ "eval_steps_per_second": 6.011,
159
+ "step": 10
160
  }
161
  ],
162
  "logging_steps": 1,
 
171
  "should_evaluate": false,
172
  "should_log": false,
173
  "should_save": true,
174
+ "should_training_stop": true
175
  },
176
  "attributes": {}
177
  }
178
  },
179
+ "total_flos": 628058829619200.0,
180
  "train_batch_size": 7,
181
  "trial_name": null,
182
  "trial_params": null