MingleiShi commited on
Commit
5c08f6f
·
verified ·
1 Parent(s): 878fca0

Upload folder using huggingface_hub

Browse files
pre-trained/autoencoder/svg_autoencoder-P-stage1.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1083c1e2ee4cabc24c0a7a51eb98a3a8177589658b01fe691048c3532f18680f
3
+ size 357108579
pre-trained/autoencoder/svg_autoencoder-P-stage2.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:457124c0a92c1c4b976b7e2a6f67cb8bba19e83ac14ae1f70de502337535b117
3
+ size 357108579
pre-trained/autoencoder/svg_autoencoder-P-stage3.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23f8f42e2384488563a90ad23691e17bee8e33d18bb6ce560a7ed108a5424285
3
+ size 357108579
pre-trained/autoencoder/svg_autoencoder-R-stage1.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60cc3612d173d57338612f30edb71043f4a009aeb6f8ea9f5ecbd2a0d1de0329
3
+ size 445917171
pre-trained/autoencoder/svg_autoencoder-R-stage2.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed64535c7809db4745c979fda99691b141b37a3a203ca2dffd325f343130a964
3
+ size 445917171
pre-trained/autoencoder/svg_autoencoder-R-stage3.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f232b8d5feac07287e225b1d1b6a2ca54226858409834a838cf397df52b592df
3
+ size 445917171
pre-trained/autoencoder/svg_autoencoder_P_stage1_256.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ training:
2
+ logdir: "logs"
3
+ scale_lr: false
4
+ accumulate_grad_batches: 1
5
+ save_every_n_train_steps: 20000
6
+ save_every_n_epochs: 1
7
+
8
+ trainer:
9
+ devices: 8
10
+ num_nodes: 1
11
+ strategy: "ddp_find_unused_parameters_true"
12
+ accelerator: "gpu"
13
+ max_epochs: 1000
14
+ precision: 32
15
+
16
+ # Model configuration
17
+ model:
18
+ init_weight: pre-trained/autoencoder/svg_autoencoder-P-stage1.ckpt
19
+ base_learning_rate: 1.0e-04
20
+ target: ldm.models.dinov3_decoder_native_resolution.DinoDecoder
21
+ params:
22
+ ckpt_path: pre-trained/autoencoder/svg_autoencoder-P-stage1.ckpt
23
+ embed_dim: 32
24
+ lossconfig:
25
+ target: "ldm.modules.losses.LPIPSWithDiscriminatorDecoder"
26
+ params:
27
+ disc_start: 5000
28
+ disc_weight: 0.5
29
+ ddconfig:
30
+ double_z: true
31
+ z_channels: 384
32
+ resolution: 256
33
+ in_channels: 3
34
+ out_ch: 3
35
+ ch: 128
36
+ ch_mult:
37
+ - 1
38
+ - 1
39
+ - 2
40
+ - 2
41
+ - 4
42
+ num_res_blocks: 2
43
+ attn_resolutions: []
44
+ dropout: 0.0
45
+
46
+ dinoconfig:
47
+ dinov3_location: dinov3
48
+ model_name: dinov3_vits16plus
49
+ weights: pre-trained/dinov3_vits16plus_pretrain_lvd1689m-4057cbaa.pth
50
+ extra_vit_config: null
51
+
52
+
pre-trained/autoencoder/svg_autoencoder_P_stage2_512.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ training:
2
+ logdir: "logs"
3
+ scale_lr: false
4
+ accumulate_grad_batches: 1
5
+ save_every_n_train_steps: 20000
6
+ save_every_n_epochs: 1
7
+
8
+ trainer:
9
+ devices: 8
10
+ num_nodes: 1
11
+ strategy: "ddp_find_unused_parameters_true"
12
+ accelerator: "gpu"
13
+ max_epochs: 1000
14
+ precision: 32
15
+
16
+ # Model configuration
17
+ model:
18
+ init_weight: pre-trained/autoencoder/svg_autoencoder-P-stage2.ckpt
19
+ base_learning_rate: 1.0e-04
20
+ target: ldm.models.dinov3_decoder_native_resolution.DinoDecoder
21
+ params:
22
+ ckpt_path: pre-trained/autoencoder/svg_autoencoder-P-stage2.ckpt
23
+ embed_dim: 32
24
+ lossconfig:
25
+ target: "ldm.modules.losses.LPIPSWithDiscriminatorDecoder"
26
+ params:
27
+ disc_start: 5000
28
+ disc_weight: 0.5
29
+ ddconfig:
30
+ double_z: true
31
+ z_channels: 384
32
+ resolution: 256
33
+ in_channels: 3
34
+ out_ch: 3
35
+ ch: 128
36
+ ch_mult:
37
+ - 1
38
+ - 1
39
+ - 2
40
+ - 2
41
+ - 4
42
+ num_res_blocks: 2
43
+ attn_resolutions: []
44
+ dropout: 0.0
45
+
46
+ dinoconfig:
47
+ dinov3_location: dinov3
48
+ model_name: dinov3_vits16plus
49
+ weights: pre-trained/dinov3_vits16plus_pretrain_lvd1689m-4057cbaa.pth
50
+ extra_vit_config: null
51
+
52
+
pre-trained/autoencoder/svg_autoencoder_P_stage3_1024.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ training:
2
+ logdir: "logs"
3
+ scale_lr: false
4
+ accumulate_grad_batches: 1
5
+ save_every_n_train_steps: 20000
6
+ save_every_n_epochs: 1
7
+
8
+ trainer:
9
+ devices: 8
10
+ num_nodes: 1
11
+ strategy: "ddp_find_unused_parameters_true"
12
+ accelerator: "gpu"
13
+ max_epochs: 1000
14
+ precision: 32
15
+
16
+ # Model configuration
17
+ model:
18
+ init_weight: pre-trained/autoencoder/svg_autoencoder-P-stage3.ckpt
19
+ base_learning_rate: 1.0e-04
20
+ target: ldm.models.dinov3_decoder_native_resolution.DinoDecoder
21
+ params:
22
+ ckpt_path: pre-trained/autoencoder/svg_autoencoder-P-stage3.ckpt
23
+ embed_dim: 32
24
+ lossconfig:
25
+ target: "ldm.modules.losses.LPIPSWithDiscriminatorDecoder"
26
+ params:
27
+ disc_start: 5000
28
+ disc_weight: 0.5
29
+ ddconfig:
30
+ double_z: true
31
+ z_channels: 384
32
+ resolution: 256
33
+ in_channels: 3
34
+ out_ch: 3
35
+ ch: 128
36
+ ch_mult:
37
+ - 1
38
+ - 1
39
+ - 2
40
+ - 2
41
+ - 4
42
+ num_res_blocks: 2
43
+ attn_resolutions: []
44
+ dropout: 0.0
45
+
46
+ dinoconfig:
47
+ dinov3_location: dinov3
48
+ model_name: dinov3_vits16plus
49
+ weights: pre-trained/dinov3_vits16plus_pretrain_lvd1689m-4057cbaa.pth
50
+ extra_vit_config: null
51
+
52
+
pre-trained/autoencoder/svg_autoencoder_R_stage1_256.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ training:
2
+ logdir: "logs"
3
+ scale_lr: false
4
+ accumulate_grad_batches: 1
5
+ save_every_n_train_steps: 20000
6
+ save_every_n_epochs: 1
7
+
8
+ trainer:
9
+ devices: 8
10
+ num_nodes: 1
11
+ strategy: "ddp_find_unused_parameters_true"
12
+ accelerator: "gpu"
13
+ max_epochs: 1000
14
+ precision: 32
15
+
16
+ # Model configuration
17
+ model:
18
+ init_weight: pre-trained/autoencoder/svg_autoencoder-R-stage1.ckpt
19
+ base_learning_rate: 1.0e-04
20
+ target: ldm.models.dinov3_decoder_native_resolution.DinoDecoder
21
+ params:
22
+ ckpt_path: pre-trained/autoencoder/svg_autoencoder-R-stage1.ckpt
23
+ embed_dim: 32
24
+ lossconfig:
25
+ target: "ldm.modules.losses.LPIPSWithDiscriminatorDecoder"
26
+ params:
27
+ disc_start: 5000
28
+ disc_weight: 0.5
29
+ ddconfig:
30
+ double_z: true
31
+ z_channels: 384
32
+ resolution: 256
33
+ in_channels: 3
34
+ out_ch: 3
35
+ ch: 128
36
+ ch_mult:
37
+ - 1
38
+ - 1
39
+ - 2
40
+ - 2
41
+ - 4
42
+ num_res_blocks: 2
43
+ attn_resolutions: []
44
+ dropout: 0.0
45
+
46
+ dinoconfig:
47
+ dinov3_location: dinov3
48
+ model_name: dinov3_vits16plus
49
+ weights: pre-trained/dinov3_vits16plus_pretrain_lvd1689m-4057cbaa.pth
50
+ extra_vit_config: null
51
+
52
+
pre-trained/autoencoder/svg_autoencoder_R_stage2_512.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ training:
2
+ logdir: "logs"
3
+ scale_lr: false
4
+ accumulate_grad_batches: 1
5
+ save_every_n_train_steps: 20000
6
+ save_every_n_epochs: 1
7
+
8
+ trainer:
9
+ devices: 8
10
+ num_nodes: 1
11
+ strategy: "ddp_find_unused_parameters_true"
12
+ accelerator: "gpu"
13
+ max_epochs: 1000
14
+ precision: 32
15
+
16
+ # Model configuration
17
+ model:
18
+ init_weight: pre-trained/autoencoder/svg_autoencoder-R-stage2.ckpt
19
+ base_learning_rate: 1.0e-04
20
+ target: ldm.models.dinov3_decoder_native_resolution.DinoDecoder
21
+ params:
22
+ ckpt_path: pre-trained/autoencoder/svg_autoencoder-R-stage2.ckpt
23
+ embed_dim: 32
24
+ lossconfig:
25
+ target: "ldm.modules.losses.LPIPSWithDiscriminatorDecoder"
26
+ params:
27
+ disc_start: 5000
28
+ disc_weight: 0.5
29
+ ddconfig:
30
+ double_z: true
31
+ z_channels: 384
32
+ resolution: 256
33
+ in_channels: 3
34
+ out_ch: 3
35
+ ch: 128
36
+ ch_mult:
37
+ - 1
38
+ - 1
39
+ - 2
40
+ - 2
41
+ - 4
42
+ num_res_blocks: 2
43
+ attn_resolutions: []
44
+ dropout: 0.0
45
+
46
+ dinoconfig:
47
+ dinov3_location: dinov3
48
+ model_name: dinov3_vits16plus
49
+ weights: pre-trained/dinov3_vits16plus_pretrain_lvd1689m-4057cbaa.pth
50
+ extra_vit_config: null
51
+
52
+
pre-trained/autoencoder/svg_autoencoder_R_stage3_1024.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ training:
2
+ logdir: "logs"
3
+ scale_lr: false
4
+ accumulate_grad_batches: 1
5
+ save_every_n_train_steps: 20000
6
+ save_every_n_epochs: 1
7
+
8
+ trainer:
9
+ devices: 8
10
+ num_nodes: 1
11
+ strategy: "ddp_find_unused_parameters_true"
12
+ accelerator: "gpu"
13
+ max_epochs: 1000
14
+ precision: 32
15
+
16
+ # Model configuration
17
+ model:
18
+ init_weight: pre-trained/autoencoder/svg_autoencoder-R-stage3.ckpt
19
+ base_learning_rate: 1.0e-04
20
+ target: ldm.models.dinov3_decoder_native_resolution.DinoDecoder
21
+ params:
22
+ ckpt_path: pre-trained/autoencoder/svg_autoencoder-R-stage3.ckpt
23
+ embed_dim: 32
24
+ lossconfig:
25
+ target: "ldm.modules.losses.LPIPSWithDiscriminatorDecoder"
26
+ params:
27
+ disc_start: 5000
28
+ disc_weight: 0.5
29
+ ddconfig:
30
+ double_z: true
31
+ z_channels: 384
32
+ resolution: 256
33
+ in_channels: 3
34
+ out_ch: 3
35
+ ch: 128
36
+ ch_mult:
37
+ - 1
38
+ - 1
39
+ - 2
40
+ - 2
41
+ - 4
42
+ num_res_blocks: 2
43
+ attn_resolutions: []
44
+ dropout: 0.0
45
+
46
+ dinoconfig:
47
+ dinov3_location: dinov3
48
+ model_name: dinov3_vits16plus
49
+ weights: pre-trained/dinov3_vits16plus_pretrain_lvd1689m-4057cbaa.pth
50
+ extra_vit_config: null
51
+
52
+