pcuenq HF Staff commited on
Commit
98fd8fb
·
1 Parent(s): 658bc95

Add configuration and weights.

Browse files

They were just copied from their original location.

config_laion_text_cond_latent_upscaler_2.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "type": "image_v1",
4
+ "input_channels": 4,
5
+ "input_size": [48, 48],
6
+ "patch_size": 1,
7
+ "mapping_out": 768,
8
+ "mapping_cond_dim": 896,
9
+ "unet_cond_dim": 4,
10
+ "depths": [4, 4, 4, 4],
11
+ "channels": [384, 384, 768, 768],
12
+ "self_attn_depths": [false, false, false, true],
13
+ "cross_attn_depths": [false, true, true, true],
14
+ "cross_cond_dim": 768,
15
+ "has_variance": true,
16
+ "dropout_rate": 0.0,
17
+ "augment_prob": 0.0,
18
+ "augment_wrapper": false,
19
+ "sigma_data": 1.0,
20
+ "sigma_min": 1e-2,
21
+ "sigma_max": 20,
22
+ "sigma_sample_density": {
23
+ "type": "lognormal",
24
+ "mean": -0.5,
25
+ "std": 1.2
26
+ }
27
+ },
28
+ "dataset": {
29
+ "type": "webdataset",
30
+ "location": "s3://s-datasets/laion-high-resolution/{00000..17535}.tar"
31
+ },
32
+ "optimizer": {
33
+ "type": "adamw",
34
+ "lr": 1.5e-4,
35
+ "betas": [0.95, 0.999],
36
+ "eps": 1e-6,
37
+ "weight_decay": 1e-3
38
+ },
39
+ "lr_sched": {
40
+ "type": "inverse",
41
+ "inv_gamma": 25000.0,
42
+ "power": 1.0,
43
+ "warmup": 0.99
44
+ },
45
+ "ema_sched": {
46
+ "type": "inverse",
47
+ "power": 0.7083,
48
+ "max_value": 0.9999
49
+ }
50
+ }
laion_text_cond_latent_upscaler_2_1_00470000_slim.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70d0fa9c6abb0061322647f9161d6829641a84cf59f9fbf65b4585b4aabdc610
3
+ size 1490493829