leffff commited on
Commit
b674477
·
verified ·
1 Parent(s): dbe90a4

Upload folder using huggingface_hub

Browse files
.ipynb_checkpoints/model_index-checkpoint.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Kandinsky5T2VPipeline",
3
+ "_diffusers_version": "0.33.0.dev0",
4
+ "scheduler": [
5
+ "diffusers",
6
+ "FlowMatchEulerDiscreteScheduler"
7
+ ],
8
+ "text_encoder": [
9
+ "transformers",
10
+ "Qwen2_5_VLForConditionalGeneration"
11
+ ],
12
+ "tokenizer": [
13
+ "transformers",
14
+ "Qwen2VLProcessor"
15
+ ],
16
+ "text_encoder_2": [
17
+ "transformers",
18
+ "CLIPTextModel"
19
+ ],
20
+ "tokenizer_2": [
21
+ "transformers",
22
+ "CLIPTokenizer"
23
+ ],
24
+ "transformer": [
25
+ "diffusers",
26
+ "Kandinsky5Transformer3DModel"
27
+ ],
28
+ "vae": [
29
+ "diffusers",
30
+ "AutoencoderKLHunyuanVideo"
31
+ ]
32
+ }
transformer/.ipynb_checkpoints/config-checkpoint.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Kandinsky5Transformer3DModel",
3
+ "_diffusers_version": "0.33.0.dev0",
4
+ "in_visual_dim": 16,
5
+ "out_visual_dim": 16,
6
+ "time_dim": 1024,
7
+ "patch_size": [1, 2, 2],
8
+ "model_dim": 4096,
9
+ "ff_dim": 16384,
10
+ "num_text_blocks": 4,
11
+ "num_visual_blocks": 60,
12
+ "axes_dims": [32, 48, 48],
13
+ "visual_cond": true,
14
+ "in_text_dim": 3584,
15
+ "in_text_dim2": 768,
16
+ "attention_type": "nabla",
17
+ "attention_chunk": false,
18
+ "attention_causal": false,
19
+ "attention_local": false,
20
+ "attention_glob": false,
21
+ "attention_window": 3,
22
+ "attention_method": "topcdf",
23
+ "attention_P": 0.8,
24
+ "attention_add_sta": true,
25
+ "attention_wT": 11,
26
+ "attention_wW": 3,
27
+ "attention_wH": 3
28
+ }
transformer/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Kandinsky5Transformer3DModel",
3
+ "_diffusers_version": "0.33.0.dev0",
4
+ "in_visual_dim": 16,
5
+ "out_visual_dim": 16,
6
+ "time_dim": 1024,
7
+ "patch_size": [1, 2, 2],
8
+ "model_dim": 4096,
9
+ "ff_dim": 16384,
10
+ "num_text_blocks": 4,
11
+ "num_visual_blocks": 60,
12
+ "axes_dims": [32, 48, 48],
13
+ "visual_cond": true,
14
+ "in_text_dim": 3584,
15
+ "in_text_dim2": 768,
16
+ "attention_type": "nabla",
17
+ "attention_causal": false,
18
+ "attention_local": false,
19
+ "attention_glob": false,
20
+ "attention_window": 3,
21
+ "attention_method": "topcdf",
22
+ "attention_P": 0.8,
23
+ "attention_add_sta": true,
24
+ "attention_wT": 11,
25
+ "attention_wW": 3,
26
+ "attention_wH": 3
27
+ }
transformer/config.txt ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Kandinsky5Transformer3DModel",
3
+ "_diffusers_version": "0.33.0.dev0",
4
+ "in_visual_dim": 16,
5
+ "out_visual_dim": 16,
6
+ "time_dim": 1024,
7
+ "patch_size": [1, 2, 2],
8
+ "model_dim": 4096,
9
+ "ff_dim": 16384,
10
+ "num_text_blocks": 4,
11
+ "num_visual_blocks": 60,
12
+ "axes_dims": [32, 48, 48],
13
+ "visual_cond": true,
14
+ "in_text_dim": 3584,
15
+ "in_text_dim2": 768,
16
+ "attention_type": "nabla",
17
+ "attention_causal": false,
18
+ "attention_local": false,
19
+ "attention_glob": false,
20
+ "attention_window": 3,
21
+ "attention_method": "topcdf",
22
+ "attention_P": 0.8,
23
+ "attention_add_sta": true,
24
+ "attention_wT": 11,
25
+ "attention_wW": 3,
26
+ "attention_wH": 3
27
+ }
transformer/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e167e9dfd4e14a2fb76a4ef3d2571df04a0d219f2f6e9121804b0dbcf5ad17c
3
+ size 77245505744