justuswill commited on
Commit
60749be
·
verified ·
1 Parent(s): 1e9f8dd

Upload 8 files

Browse files
checkpoints/uqdm-big/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44d51e7933a76db667bd2c308ced7b93248fe76c02a9d4da1652bae58991f7d2
3
+ size 3803823283
checkpoints/uqdm-big/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "restore_ckpt": "checkpoint.pt",
3
+ "model": {
4
+ "n_timesteps": 4,
5
+ "prior_type": "logistic",
6
+ "learned_prior_scale": true,
7
+ "base_prior_scale": "forward_kernel",
8
+ "noise_schedule": "learned_linear",
9
+ "fix_gamma_max": true,
10
+ "gamma_min": -12.3,
11
+ "gamma_max": 6.0,
12
+ "attention_everywhere": false,
13
+ "use_fourier_features": true,
14
+ "ema_rate": 0.9999,
15
+ "n_attention_heads": 1,
16
+ "n_channels": 3,
17
+ "vocab_size": 256,
18
+ "embedding_dim": 256,
19
+ "n_blocks": 64,
20
+ "norm_groups": 32,
21
+ "dropout_prob": 0.0
22
+ },
23
+ "training": {
24
+ "n_steps": 500000,
25
+ "log_metrics_every_steps": 1000,
26
+ "checkpoint_every_steps": 10000,
27
+ "eval_every_steps": 10000,
28
+ "eval_steps_to_run": 50
29
+ },
30
+ "optim": {
31
+ "weight_decay": 0,
32
+ "beta1": 0.9,
33
+ "eps": 1e-08,
34
+ "warmup": 1000,
35
+ "lr": 0.0002,
36
+ "grad_clip_norm": 1.0
37
+ },
38
+ "data": {
39
+ "shuffle": false,
40
+ "pin_memory": true,
41
+ "batch_size": 2,
42
+ "num_workers": 1,
43
+ "data_spec": "imagenet"
44
+ }
45
+ }
checkpoints/uqdm-medium/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4326f9460bd41311a340e2c7b2decce04c2412a093cab592d2572fba05622c7d
3
+ size 1955363955
checkpoints/uqdm-medium/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "restore_ckpt": "checkpoint.pt",
3
+ "model": {
4
+ "n_timesteps": 4,
5
+ "prior_type": "logistic",
6
+ "learned_prior_scale": true,
7
+ "base_prior_scale": "forward_kernel",
8
+ "noise_schedule": "learned_linear",
9
+ "fix_gamma_max": true,
10
+ "gamma_min": -12.3,
11
+ "gamma_max": 6.0,
12
+ "attention_everywhere": false,
13
+ "use_fourier_features": true,
14
+ "ema_rate": 0.9999,
15
+ "n_attention_heads": 1,
16
+ "n_channels": 3,
17
+ "vocab_size": 256,
18
+ "embedding_dim": 256,
19
+ "n_blocks": 32,
20
+ "norm_groups": 32,
21
+ "dropout_prob": 0.0
22
+ },
23
+ "training": {
24
+ "n_steps": 500000,
25
+ "log_metrics_every_steps": 1000,
26
+ "checkpoint_every_steps": 10000,
27
+ "eval_every_steps": 10000,
28
+ "eval_steps_to_run": 50
29
+ },
30
+ "optim": {
31
+ "weight_decay": 0,
32
+ "beta1": 0.9,
33
+ "eps": 1e-08,
34
+ "warmup": 1000,
35
+ "lr": 0.0002,
36
+ "grad_clip_norm": 1.0
37
+ },
38
+ "data": {
39
+ "shuffle": false,
40
+ "pin_memory": true,
41
+ "batch_size": 2,
42
+ "num_workers": 1,
43
+ "data_spec": "imagenet"
44
+ }
45
+ }
checkpoints/uqdm-small/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:339fa93762cb34f253118c3f8e058f2ecb8aacea93746b94c8e7182779a4f497
3
+ size 36126963
checkpoints/uqdm-small/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "restore_ckpt": "checkpoint.pt",
3
+ "model": {
4
+ "n_timesteps": 4,
5
+ "prior_type": "logistic",
6
+ "learned_prior_scale": true,
7
+ "base_prior_scale": "forward_kernel",
8
+ "noise_schedule": "learned_linear",
9
+ "fix_gamma_max": true,
10
+ "gamma_min": -12.3,
11
+ "gamma_max": 6.0,
12
+ "attention_everywhere": false,
13
+ "use_fourier_features": true,
14
+ "ema_rate": 0.9999,
15
+ "n_attention_heads": 1,
16
+ "n_channels": 3,
17
+ "vocab_size": 256,
18
+ "embedding_dim": 64,
19
+ "n_blocks": 8,
20
+ "norm_groups": 16,
21
+ "dropout_prob": 0.0
22
+ },
23
+ "training": {
24
+ "n_steps": 500000,
25
+ "log_metrics_every_steps": 1000,
26
+ "checkpoint_every_steps": 10000,
27
+ "eval_every_steps": 10000,
28
+ "eval_steps_to_run": 50
29
+ },
30
+ "optim": {
31
+ "weight_decay": 0,
32
+ "beta1": 0.9,
33
+ "eps": 1e-08,
34
+ "warmup": 1000,
35
+ "lr": 0.0002,
36
+ "grad_clip_norm": 1.0
37
+ },
38
+ "data": {
39
+ "shuffle": false,
40
+ "pin_memory": true,
41
+ "batch_size": 2,
42
+ "num_workers": 1,
43
+ "data_spec": "imagenet"
44
+ }
45
+ }
checkpoints/uqdm-tiny/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52644fcb40135471a2d2b526ea67ccb888425b472e8f889117fbee0ede428637
3
+ size 2781146
checkpoints/uqdm-tiny/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "restore_ckpt": "checkpoint.pt",
3
+ "model": {
4
+ "n_timesteps": 4,
5
+ "prior_type": "logistic",
6
+ "learned_prior_scale": true,
7
+ "base_prior_scale": "forward_kernel",
8
+ "noise_schedule": "learned_linear",
9
+ "fix_gamma_max": true,
10
+ "gamma_min": -12.3,
11
+ "gamma_max": 6.0,
12
+ "attention_everywhere": false,
13
+ "use_fourier_features": true,
14
+ "ema_rate": 0.9999,
15
+ "n_attention_heads": 1,
16
+ "n_channels": 3,
17
+ "vocab_size": 256,
18
+ "embedding_dim": 32,
19
+ "n_blocks": 1,
20
+ "norm_groups": 8,
21
+ "dropout_prob": 0.0
22
+ },
23
+ "training": {
24
+ "n_steps": 500000,
25
+ "log_metrics_every_steps": 1000,
26
+ "checkpoint_every_steps": 10000,
27
+ "eval_every_steps": 10000,
28
+ "eval_steps_to_run": 50
29
+ },
30
+ "optim": {
31
+ "weight_decay": 0,
32
+ "beta1": 0.9,
33
+ "eps": 1e-08,
34
+ "warmup": 1000,
35
+ "lr": 0.0002,
36
+ "grad_clip_norm": 1.0
37
+ },
38
+ "data": {
39
+ "shuffle": false,
40
+ "pin_memory": true,
41
+ "batch_size": 2,
42
+ "num_workers": 1,
43
+ "data_spec": "imagenet"
44
+ }
45
+ }