rootxhacker commited on
Commit
75a19b2
·
verified ·
1 Parent(s): 9b5a9e8

Training in progress, step 12000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66d99b0cb397502b025a6c7417fc75e69557e2db9d4b2101658c2730782a54e5
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da26990df234d94448e406f402bbaab14595948f6b2bb6aaf6233e11502b39c8
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eacebe9389e2329614f557172e304528d2b5b66f68c1b3c5e68453452e430f51
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fd8f2ea0f8f3f2b6317bf012fe6c554f0ab9b95b3c3bc864dd5cae3b6eef992
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a032319df626b2b110bd852b78fc23d492586d6fccaeb0841874a11adbfaea49
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70a89db0f278efb20b8c741665f32f39c512b7877cd595dd7f15cfb9e97916f3
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a16095fe7c7a01fb5cdfa33bc3b7f9026f465c5d2d551f188966908ce9a84c9
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1696a7b978a5af1f9de4c1c4ce7f56376b2325426ded966dd6d734962f026c10
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2124619b96ef0628e278d9139421c199b6678f87b3a0cfd00afb352c22439c91
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f076f7a63b00dbdf4d7a2eb7609297124339e4b64abf631b5119fa735b47e6a
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:129303cede08862e45aff723e13523f2863b1a8c5dd6144e719bcbf05975af10
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7a1d25be56e05cb4cbafc0981ade0df33b24e8c1de440cdbae85f1fea41cdfa
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 11500,
3
- "best_metric": 1.3371928930282593,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-11500",
5
- "epoch": 0.8845473425121144,
6
  "eval_steps": 250,
7
- "global_step": 11500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1986,6 +1986,92 @@
1986
  "eval_samples_per_second": 59.282,
1987
  "eval_steps_per_second": 14.821,
1988
  "step": 11500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1989
  }
1990
  ],
1991
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 12000,
3
+ "best_metric": -30.337785720825195,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-12000",
5
+ "epoch": 0.9230059226213368,
6
  "eval_steps": 250,
7
+ "global_step": 12000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1986
  "eval_samples_per_second": 59.282,
1987
  "eval_steps_per_second": 14.821,
1988
  "step": 11500
1989
+ },
1990
+ {
1991
+ "epoch": 0.8883932005230367,
1992
+ "grad_norm": 2.3821170330047607,
1993
+ "learning_rate": 0.00014264343038204814,
1994
+ "loss": 1.4094,
1995
+ "step": 11550
1996
+ },
1997
+ {
1998
+ "epoch": 0.8922390585339589,
1999
+ "grad_norm": 1.0703688859939575,
2000
+ "learning_rate": 0.00014238371036023168,
2001
+ "loss": 1.3967,
2002
+ "step": 11600
2003
+ },
2004
+ {
2005
+ "epoch": 0.8960849165448812,
2006
+ "grad_norm": 0.9679076075553894,
2007
+ "learning_rate": 0.0001421239903384152,
2008
+ "loss": 1.3855,
2009
+ "step": 11650
2010
+ },
2011
+ {
2012
+ "epoch": 0.8999307745558034,
2013
+ "grad_norm": 1.2558847665786743,
2014
+ "learning_rate": 0.00014186427031659873,
2015
+ "loss": 1.3501,
2016
+ "step": 11700
2017
+ },
2018
+ {
2019
+ "epoch": 0.9037766325667257,
2020
+ "grad_norm": 2.6319808959960938,
2021
+ "learning_rate": 0.00014160455029478224,
2022
+ "loss": -15.5666,
2023
+ "step": 11750
2024
+ },
2025
+ {
2026
+ "epoch": 0.9037766325667257,
2027
+ "eval_loss": -26.468090057373047,
2028
+ "eval_runtime": 16.5143,
2029
+ "eval_samples_per_second": 60.554,
2030
+ "eval_steps_per_second": 15.138,
2031
+ "step": 11750
2032
+ },
2033
+ {
2034
+ "epoch": 0.9076224905776479,
2035
+ "grad_norm": 2.3091208934783936,
2036
+ "learning_rate": 0.00014134483027296576,
2037
+ "loss": -28.8843,
2038
+ "step": 11800
2039
+ },
2040
+ {
2041
+ "epoch": 0.9114683485885701,
2042
+ "grad_norm": 1.527166724205017,
2043
+ "learning_rate": 0.00014108511025114927,
2044
+ "loss": -30.1099,
2045
+ "step": 11850
2046
+ },
2047
+ {
2048
+ "epoch": 0.9153142065994924,
2049
+ "grad_norm": 3.9258713722229004,
2050
+ "learning_rate": 0.00014082539022933278,
2051
+ "loss": -30.2295,
2052
+ "step": 11900
2053
+ },
2054
+ {
2055
+ "epoch": 0.9191600646104146,
2056
+ "grad_norm": 1.618696928024292,
2057
+ "learning_rate": 0.0001405656702075163,
2058
+ "loss": -30.2778,
2059
+ "step": 11950
2060
+ },
2061
+ {
2062
+ "epoch": 0.9230059226213368,
2063
+ "grad_norm": 1.5081305503845215,
2064
+ "learning_rate": 0.00014030595018569984,
2065
+ "loss": -30.3095,
2066
+ "step": 12000
2067
+ },
2068
+ {
2069
+ "epoch": 0.9230059226213368,
2070
+ "eval_loss": -30.337785720825195,
2071
+ "eval_runtime": 16.3273,
2072
+ "eval_samples_per_second": 61.247,
2073
+ "eval_steps_per_second": 15.312,
2074
+ "step": 12000
2075
  }
2076
  ],
2077
  "logging_steps": 50,