rootxhacker commited on
Commit
0b40601
·
verified ·
1 Parent(s): c8cbecd

Training in progress, step 12500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da26990df234d94448e406f402bbaab14595948f6b2bb6aaf6233e11502b39c8
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a623f2e366d56a653f11e3f01819889dc1645d4d1136c5b510bf0c1c92dab66
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4fd8f2ea0f8f3f2b6317bf012fe6c554f0ab9b95b3c3bc864dd5cae3b6eef992
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6015f4fec963cdcdbd9534ce342304afb8094ad5f761558a674ddf3068ab0335
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70a89db0f278efb20b8c741665f32f39c512b7877cd595dd7f15cfb9e97916f3
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f314be3a1ba2ce8f5f725cf645fa00681145b061c07cf8ad59e1e13356924cb
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1696a7b978a5af1f9de4c1c4ce7f56376b2325426ded966dd6d734962f026c10
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:082b149e63f0f47e9e60d37040a9d015e2fa6a3d80388f6fa7d859aa3870fd21
3
+ size 14308
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f076f7a63b00dbdf4d7a2eb7609297124339e4b64abf631b5119fa735b47e6a
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0678e6da71fe39cdc83977acb14900d45184a734b31b0b8a20ead5262367273c
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7a1d25be56e05cb4cbafc0981ade0df33b24e8c1de440cdbae85f1fea41cdfa
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7680f728f95fcd934cd32ef342f02b8f2b507472a27e9e8de734cc5468e1428
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 12000,
3
- "best_metric": -30.337785720825195,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-12000",
5
- "epoch": 0.9230059226213368,
6
  "eval_steps": 250,
7
- "global_step": 12000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2072,6 +2072,92 @@
2072
  "eval_samples_per_second": 61.247,
2073
  "eval_steps_per_second": 15.312,
2074
  "step": 12000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2075
  }
2076
  ],
2077
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 12500,
3
+ "best_metric": -30.455406188964844,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-12500",
5
+ "epoch": 0.9614645027305592,
6
  "eval_steps": 250,
7
+ "global_step": 12500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2072
  "eval_samples_per_second": 61.247,
2073
  "eval_steps_per_second": 15.312,
2074
  "step": 12000
2075
+ },
2076
+ {
2077
+ "epoch": 0.926851780632259,
2078
+ "grad_norm": 1.3591136932373047,
2079
+ "learning_rate": 0.00014004623016388335,
2080
+ "loss": -30.3233,
2081
+ "step": 12050
2082
+ },
2083
+ {
2084
+ "epoch": 0.9306976386431813,
2085
+ "grad_norm": 1.8113991022109985,
2086
+ "learning_rate": 0.00013978651014206686,
2087
+ "loss": -30.3411,
2088
+ "step": 12100
2089
+ },
2090
+ {
2091
+ "epoch": 0.9345434966541035,
2092
+ "grad_norm": 2.9664206504821777,
2093
+ "learning_rate": 0.00013952679012025037,
2094
+ "loss": -30.3763,
2095
+ "step": 12150
2096
+ },
2097
+ {
2098
+ "epoch": 0.9383893546650258,
2099
+ "grad_norm": 4.572649955749512,
2100
+ "learning_rate": 0.0001392670700984339,
2101
+ "loss": -30.3894,
2102
+ "step": 12200
2103
+ },
2104
+ {
2105
+ "epoch": 0.942235212675948,
2106
+ "grad_norm": 0.7282872200012207,
2107
+ "learning_rate": 0.0001390073500766174,
2108
+ "loss": -30.3943,
2109
+ "step": 12250
2110
+ },
2111
+ {
2112
+ "epoch": 0.942235212675948,
2113
+ "eval_loss": -30.41870880126953,
2114
+ "eval_runtime": 16.3228,
2115
+ "eval_samples_per_second": 61.264,
2116
+ "eval_steps_per_second": 15.316,
2117
+ "step": 12250
2118
+ },
2119
+ {
2120
+ "epoch": 0.9460810706868702,
2121
+ "grad_norm": 0.5192355513572693,
2122
+ "learning_rate": 0.00013874763005480094,
2123
+ "loss": -30.4192,
2124
+ "step": 12300
2125
+ },
2126
+ {
2127
+ "epoch": 0.9499269286977925,
2128
+ "grad_norm": 0.4913904070854187,
2129
+ "learning_rate": 0.00013848791003298445,
2130
+ "loss": -30.4326,
2131
+ "step": 12350
2132
+ },
2133
+ {
2134
+ "epoch": 0.9537727867087147,
2135
+ "grad_norm": 0.36102983355522156,
2136
+ "learning_rate": 0.00013822819001116797,
2137
+ "loss": -30.4425,
2138
+ "step": 12400
2139
+ },
2140
+ {
2141
+ "epoch": 0.957618644719637,
2142
+ "grad_norm": 0.1661052107810974,
2143
+ "learning_rate": 0.00013796846998935148,
2144
+ "loss": -30.4485,
2145
+ "step": 12450
2146
+ },
2147
+ {
2148
+ "epoch": 0.9614645027305592,
2149
+ "grad_norm": 0.43040284514427185,
2150
+ "learning_rate": 0.000137708749967535,
2151
+ "loss": -30.4529,
2152
+ "step": 12500
2153
+ },
2154
+ {
2155
+ "epoch": 0.9614645027305592,
2156
+ "eval_loss": -30.455406188964844,
2157
+ "eval_runtime": 16.2806,
2158
+ "eval_samples_per_second": 61.423,
2159
+ "eval_steps_per_second": 15.356,
2160
+ "step": 12500
2161
  }
2162
  ],
2163
  "logging_steps": 50,