lanko commited on
Commit
804a5f7
·
1 Parent(s): cf4c005

Upload folder using huggingface_hub

Browse files
v5-2/checkpoint-3600/transformer/config.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanVACETransformer3DModel",
3
+ "_diffusers_version": "0.35.1",
4
+ "_name_or_path": "/home/jingchen/links/scratch/hcplab/wangzijun/checkpoints/mopc-v5.2/checkpoint-3300/transformer",
5
+ "added_kv_proj_dim": null,
6
+ "attention_head_dim": 128,
7
+ "cross_attn_norm": true,
8
+ "eps": 1e-06,
9
+ "expert_thw": [
10
+ 21,
11
+ 60,
12
+ 104
13
+ ],
14
+ "ffn_dim": 8960,
15
+ "freq_dim": 256,
16
+ "image_dim": null,
17
+ "in_channels": 16,
18
+ "local_moe_ffn_dim": 1536,
19
+ "local_num_experts": 32,
20
+ "local_physical_classes": 23,
21
+ "local_topk": 4,
22
+ "num_attention_heads": 12,
23
+ "num_experts": 32,
24
+ "num_layers": 30,
25
+ "out_channels": 16,
26
+ "patch_size": [
27
+ 1,
28
+ 2,
29
+ 2
30
+ ],
31
+ "pos_embed_seq_len": null,
32
+ "qk_norm": "rms_norm_across_heads",
33
+ "rope_max_seq_len": 1024,
34
+ "router_temperature": 4.0,
35
+ "text_dim": 4096,
36
+ "vace_in_channels": 48,
37
+ "vace_layers": [
38
+ 0,
39
+ 2,
40
+ 4,
41
+ 6,
42
+ 8,
43
+ 10,
44
+ 12,
45
+ 14,
46
+ 16,
47
+ 18,
48
+ 20,
49
+ 22,
50
+ 24,
51
+ 26,
52
+ 28
53
+ ]
54
+ }
v5-2/checkpoint-3600/transformer/diffusion_pytorch_model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f891725fd47cdb90d3c0733216d2af5ee2c6272d53c2b1e59c0983475119f83a
3
+ size 4999723872
v5-2/checkpoint-3600/transformer/diffusion_pytorch_model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a48dddaff590b69985bea240461ae656a49e420f0456d2e14ff464318f42baa
3
+ size 2382998208
v5-2/checkpoint-3600/transformer/diffusion_pytorch_model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff