rootxhacker commited on
Commit
a3986b0
·
verified ·
1 Parent(s): 88a279e

Training in progress, step 11500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:932444defba9a5b8dd44f88b8bcb52ab659664a28676382e15a373cf36bf6034
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:951de22caec4595c4ffeef6e17bf618ff4c5009026d3bf61fbbf6c21e394a753
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f80012b6836a364fdec2135af886ce664e83e4046832b07a32b6501d8f4ef656
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94352e8877cdd376b4e0992fbcd1a31c79a5cbb8edc941287bffbb6813ad2fd9
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d7167878bddb7c0a88c66ad0b0be73885dac80e1980348d5e68d660c4a16ce9
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ce35713c6cfde8b254092a62a162b21058e7d89b73dff276ea2ee9e65249418
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42adf2998e15e06307e725ba7dcb6c52fc0a94a297e1345e89da412c057a53a6
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7733da2de8ba0d811f711ea7fdb6cdb69bbfeb6cd9a30bcfdad50f83ead3040c
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f034f97f7b7aead49a2646d657b2abae10ec15daef82d04e9386fb7172dbeddc
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d47850787d7113906407c13287aaa8fd754cf26d87c84c8d6940554fabd6e6fa
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 11000,
3
  "best_metric": 4.307990074157715,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-11000",
5
- "epoch": 0.8460887624028921,
6
  "eval_steps": 250,
7
- "global_step": 11000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1900,6 +1900,92 @@
1900
  "eval_samples_per_second": 53.273,
1901
  "eval_steps_per_second": 13.318,
1902
  "step": 11000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1903
  }
1904
  ],
1905
  "logging_steps": 50,
 
2
  "best_global_step": 11000,
3
  "best_metric": 4.307990074157715,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-11000",
5
+ "epoch": 0.8845473425121144,
6
  "eval_steps": 250,
7
+ "global_step": 11500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1900
  "eval_samples_per_second": 53.273,
1901
  "eval_steps_per_second": 13.318,
1902
  "step": 11000
1903
+ },
1904
+ {
1905
+ "epoch": 0.8499346204138143,
1906
+ "grad_norm": 5.336431503295898,
1907
+ "learning_rate": 8.36983934414176e-05,
1908
+ "loss": 4.3977,
1909
+ "step": 11050
1910
+ },
1911
+ {
1912
+ "epoch": 0.8537804784247366,
1913
+ "grad_norm": 4.175157070159912,
1914
+ "learning_rate": 8.344104996139847e-05,
1915
+ "loss": 4.4053,
1916
+ "step": 11100
1917
+ },
1918
+ {
1919
+ "epoch": 0.8576263364356588,
1920
+ "grad_norm": 4.384688377380371,
1921
+ "learning_rate": 8.318370648137934e-05,
1922
+ "loss": 4.26,
1923
+ "step": 11150
1924
+ },
1925
+ {
1926
+ "epoch": 0.8614721944465811,
1927
+ "grad_norm": 3.6022467613220215,
1928
+ "learning_rate": 8.292636300136024e-05,
1929
+ "loss": 4.2993,
1930
+ "step": 11200
1931
+ },
1932
+ {
1933
+ "epoch": 0.8653180524575033,
1934
+ "grad_norm": 4.252429485321045,
1935
+ "learning_rate": 8.266901952134112e-05,
1936
+ "loss": 4.299,
1937
+ "step": 11250
1938
+ },
1939
+ {
1940
+ "epoch": 0.8653180524575033,
1941
+ "eval_loss": 4.334308624267578,
1942
+ "eval_runtime": 18.9071,
1943
+ "eval_samples_per_second": 52.89,
1944
+ "eval_steps_per_second": 13.223,
1945
+ "step": 11250
1946
+ },
1947
+ {
1948
+ "epoch": 0.8691639104684256,
1949
+ "grad_norm": 3.4003775119781494,
1950
+ "learning_rate": 8.241167604132201e-05,
1951
+ "loss": 4.2806,
1952
+ "step": 11300
1953
+ },
1954
+ {
1955
+ "epoch": 0.8730097684793477,
1956
+ "grad_norm": 3.7436835765838623,
1957
+ "learning_rate": 8.215433256130289e-05,
1958
+ "loss": 4.2694,
1959
+ "step": 11350
1960
+ },
1961
+ {
1962
+ "epoch": 0.8768556264902699,
1963
+ "grad_norm": 2.8963701725006104,
1964
+ "learning_rate": 8.189698908128376e-05,
1965
+ "loss": 4.362,
1966
+ "step": 11400
1967
+ },
1968
+ {
1969
+ "epoch": 0.8807014845011922,
1970
+ "grad_norm": 3.3496339321136475,
1971
+ "learning_rate": 8.163964560126465e-05,
1972
+ "loss": 4.3698,
1973
+ "step": 11450
1974
+ },
1975
+ {
1976
+ "epoch": 0.8845473425121144,
1977
+ "grad_norm": 4.4007487297058105,
1978
+ "learning_rate": 8.138230212124553e-05,
1979
+ "loss": 4.2994,
1980
+ "step": 11500
1981
+ },
1982
+ {
1983
+ "epoch": 0.8845473425121144,
1984
+ "eval_loss": 4.315768241882324,
1985
+ "eval_runtime": 18.8056,
1986
+ "eval_samples_per_second": 53.176,
1987
+ "eval_steps_per_second": 13.294,
1988
+ "step": 11500
1989
  }
1990
  ],
1991
  "logging_steps": 50,