rootxhacker commited on
Commit
a69c5fa
·
verified ·
1 Parent(s): e6b4686

Training in progress, step 1000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fae9e588eaaf24844115c1eddab6702485f0e276176d9905391edb7633c50bea
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f02ec6b82a831f2902a910cf8fafe953154448b3b45acb102d57d8391c23b96
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:46cbbd66b0fad4cd21043dde397564b4a5a43cb08fd3e63a8a9d298088a80ea8
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbe39982c4bebe4e0b3f0f76afe1c485909ad223aac0a865ef49d364896c3cff
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fecf71fcdecddc7b1d484764aa9989b5c7a376940de2345dbcd4727173f1a8cc
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c326c13c75092245145edf95faedf4bbc2e6d95bbc78269d58ccc6219fd83b99
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6091cec722ce83576335fe69e6a4694ee54be33f028be0bbcfa0367b80c50012
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31bb3f2abd407f62d555fa53e113286139c6abb36d00dab486b238b592c7d71a
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad6acbbacc4296909519356cab55a9b5586a68ca85ba3af712eed9f037b1e773
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2229a15bdf42c8dbae70a65f18e33cde5c16363745edce9ae10c3d4c94d5ed8e
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df8159a5d54d37b240325feb44ba0ce415d32bac74fe814243062911256e4717
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d88837be00a1ec9be1702d7436c136fd03c05aff969a737326a957a88901e17
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 500,
3
- "best_metric": 1.7362440824508667,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-500",
5
- "epoch": 0.038458580109222366,
6
  "eval_steps": 250,
7
- "global_step": 500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -94,6 +94,92 @@
94
  "eval_samples_per_second": 59.399,
95
  "eval_steps_per_second": 14.85,
96
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97
  }
98
  ],
99
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 1000,
3
+ "best_metric": 1.6695575714111328,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-1000",
5
+ "epoch": 0.07691716021844473,
6
  "eval_steps": 250,
7
+ "global_step": 1000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
94
  "eval_samples_per_second": 59.399,
95
  "eval_steps_per_second": 14.85,
96
  "step": 500
97
+ },
98
+ {
99
+ "epoch": 0.0423044381201446,
100
+ "grad_norm": 1.872428297996521,
101
+ "learning_rate": 0.00019976105757992885,
102
+ "loss": 1.7652,
103
+ "step": 550
104
+ },
105
+ {
106
+ "epoch": 0.04615029613106684,
107
+ "grad_norm": 1.5747038125991821,
108
+ "learning_rate": 0.00019950133755811236,
109
+ "loss": 1.7437,
110
+ "step": 600
111
+ },
112
+ {
113
+ "epoch": 0.04999615414198908,
114
+ "grad_norm": 2.0243659019470215,
115
+ "learning_rate": 0.00019924161753629587,
116
+ "loss": 1.6628,
117
+ "step": 650
118
+ },
119
+ {
120
+ "epoch": 0.05384201215291132,
121
+ "grad_norm": 1.1568059921264648,
122
+ "learning_rate": 0.0001989818975144794,
123
+ "loss": 1.6787,
124
+ "step": 700
125
+ },
126
+ {
127
+ "epoch": 0.05768787016383355,
128
+ "grad_norm": 1.8709088563919067,
129
+ "learning_rate": 0.0001987221774926629,
130
+ "loss": 1.7599,
131
+ "step": 750
132
+ },
133
+ {
134
+ "epoch": 0.05768787016383355,
135
+ "eval_loss": 1.6890833377838135,
136
+ "eval_runtime": 16.858,
137
+ "eval_samples_per_second": 59.319,
138
+ "eval_steps_per_second": 14.83,
139
+ "step": 750
140
+ },
141
+ {
142
+ "epoch": 0.06153372817475579,
143
+ "grad_norm": 1.4875328540802002,
144
+ "learning_rate": 0.00019846245747084644,
145
+ "loss": 1.7187,
146
+ "step": 800
147
+ },
148
+ {
149
+ "epoch": 0.06537958618567802,
150
+ "grad_norm": 1.9822659492492676,
151
+ "learning_rate": 0.00019820273744902995,
152
+ "loss": 1.7005,
153
+ "step": 850
154
+ },
155
+ {
156
+ "epoch": 0.06922544419660026,
157
+ "grad_norm": 1.3554214239120483,
158
+ "learning_rate": 0.0001979430174272135,
159
+ "loss": 1.7001,
160
+ "step": 900
161
+ },
162
+ {
163
+ "epoch": 0.0730713022075225,
164
+ "grad_norm": 0.9904000163078308,
165
+ "learning_rate": 0.000197683297405397,
166
+ "loss": 1.6042,
167
+ "step": 950
168
+ },
169
+ {
170
+ "epoch": 0.07691716021844473,
171
+ "grad_norm": 1.4929814338684082,
172
+ "learning_rate": 0.00019742877178401683,
173
+ "loss": 1.7251,
174
+ "step": 1000
175
+ },
176
+ {
177
+ "epoch": 0.07691716021844473,
178
+ "eval_loss": 1.6695575714111328,
179
+ "eval_runtime": 16.8334,
180
+ "eval_samples_per_second": 59.406,
181
+ "eval_steps_per_second": 14.851,
182
+ "step": 1000
183
  }
184
  ],
185
  "logging_steps": 50,