besimray commited on
Commit
00ffe99
·
verified ·
1 Parent(s): eb1fc45

Training in progress, step 10, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4d2c93dc5e6fe1c8061e8407ae5586210c4af0b0962e99d588182e9b11af41d
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43da2a1eb2790fe62f293e01d8ada013a0d966df4f2ada2660b9a71af9f5002a
3
  size 45118424
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5234a04f78dd0b72e6ab48c05b348bc8f273c06f7b5ab51e050911896206e570
3
  size 23159290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39701fa216f95fd1e0171751e98aa83608c72ac9eb0c890e4ab69484314eeeed
3
  size 23159290
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a85c74230107fa7712e9fba79705e62f241775a7bdb5cde899d6d3cfa910a86
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13893f254074b0dec5e8003565583ed9154abab49baf146f6045b5d48206e343
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cda2d7da3ce8a95be6df505b84eace6ccd5aa18ffc6d1bcc9a79572045c8d78f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb578e75c11a81e85dda67a691f96ba4793a02960f1409fd3e1511aac873491a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.10526315789473684,
5
  "eval_steps": 2,
6
- "global_step": 5,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -66,6 +66,65 @@
66
  "learning_rate": 0.0001,
67
  "loss": 1.2373,
68
  "step": 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
  }
70
  ],
71
  "logging_steps": 1,
@@ -80,12 +139,12 @@
80
  "should_evaluate": false,
81
  "should_log": false,
82
  "should_save": true,
83
- "should_training_stop": false
84
  },
85
  "attributes": {}
86
  }
87
  },
88
- "total_flos": 4838612847820800.0,
89
  "train_batch_size": 10,
90
  "trial_name": null,
91
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.21052631578947367,
5
  "eval_steps": 2,
6
+ "global_step": 10,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
66
  "learning_rate": 0.0001,
67
  "loss": 1.2373,
68
  "step": 5
69
+ },
70
+ {
71
+ "epoch": 0.12631578947368421,
72
+ "grad_norm": 0.5380682349205017,
73
+ "learning_rate": 0.00012,
74
+ "loss": 1.3893,
75
+ "step": 6
76
+ },
77
+ {
78
+ "epoch": 0.12631578947368421,
79
+ "eval_loss": 1.2157351970672607,
80
+ "eval_runtime": 15.0706,
81
+ "eval_samples_per_second": 6.635,
82
+ "eval_steps_per_second": 0.664,
83
+ "step": 6
84
+ },
85
+ {
86
+ "epoch": 0.14736842105263157,
87
+ "grad_norm": 0.6841205358505249,
88
+ "learning_rate": 0.00014,
89
+ "loss": 1.3356,
90
+ "step": 7
91
+ },
92
+ {
93
+ "epoch": 0.16842105263157894,
94
+ "grad_norm": 0.4957394599914551,
95
+ "learning_rate": 0.00016,
96
+ "loss": 1.2216,
97
+ "step": 8
98
+ },
99
+ {
100
+ "epoch": 0.16842105263157894,
101
+ "eval_loss": 1.2087069749832153,
102
+ "eval_runtime": 15.1596,
103
+ "eval_samples_per_second": 6.596,
104
+ "eval_steps_per_second": 0.66,
105
+ "step": 8
106
+ },
107
+ {
108
+ "epoch": 0.18947368421052632,
109
+ "grad_norm": 0.5728506445884705,
110
+ "learning_rate": 0.00018,
111
+ "loss": 1.2016,
112
+ "step": 9
113
+ },
114
+ {
115
+ "epoch": 0.21052631578947367,
116
+ "grad_norm": 0.4859888255596161,
117
+ "learning_rate": 0.0002,
118
+ "loss": 1.3764,
119
+ "step": 10
120
+ },
121
+ {
122
+ "epoch": 0.21052631578947367,
123
+ "eval_loss": 1.1915773153305054,
124
+ "eval_runtime": 15.0938,
125
+ "eval_samples_per_second": 6.625,
126
+ "eval_steps_per_second": 0.663,
127
+ "step": 10
128
  }
129
  ],
130
  "logging_steps": 1,
 
139
  "should_evaluate": false,
140
  "should_log": false,
141
  "should_save": true,
142
+ "should_training_stop": true
143
  },
144
  "attributes": {}
145
  }
146
  },
147
+ "total_flos": 9677225695641600.0,
148
  "train_batch_size": 10,
149
  "trial_name": null,
150
  "trial_params": null