mrfatso commited on
Commit
5b64d79
·
verified ·
1 Parent(s): 6cda23c

Upload Anima preview 3 Diffusers model

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer/tokenizer.json filter=lfs diff=lfs merge=lfs -text
model_index.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AnimaPipeline",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "scheduler": [
5
+ "diffusers",
6
+ "FlowMatchEulerDiscreteScheduler"
7
+ ],
8
+ "t5_tokenizer": [
9
+ "transformers",
10
+ "T5Tokenizer"
11
+ ],
12
+ "text_conditioner": [
13
+ "anima",
14
+ "AnimaTextConditioner"
15
+ ],
16
+ "text_encoder": [
17
+ "transformers",
18
+ "Qwen3Model"
19
+ ],
20
+ "tokenizer": [
21
+ "transformers",
22
+ "Qwen2Tokenizer"
23
+ ],
24
+ "transformer": [
25
+ "diffusers",
26
+ "CosmosTransformer3DModel"
27
+ ],
28
+ "vae": [
29
+ "diffusers",
30
+ "AutoencoderKLQwenImage"
31
+ ]
32
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FlowMatchEulerDiscreteScheduler",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "base_image_seq_len": 256,
5
+ "base_shift": 0.5,
6
+ "invert_sigmas": false,
7
+ "max_image_seq_len": 4096,
8
+ "max_shift": 1.15,
9
+ "num_train_timesteps": 1000,
10
+ "shift": 3.0,
11
+ "shift_terminal": null,
12
+ "stochastic_sampling": false,
13
+ "time_shift_type": "exponential",
14
+ "use_beta_sigmas": false,
15
+ "use_dynamic_shifting": false,
16
+ "use_exponential_sigmas": false,
17
+ "use_karras_sigmas": false
18
+ }
t5_tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
t5_tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "clean_up_tokenization_spaces": true,
4
+ "eos_token": "</s>",
5
+ "extra_ids": 100,
6
+ "extra_special_tokens": [
7
+ "<extra_id_0>",
8
+ "<extra_id_1>",
9
+ "<extra_id_2>",
10
+ "<extra_id_3>",
11
+ "<extra_id_4>",
12
+ "<extra_id_5>",
13
+ "<extra_id_6>",
14
+ "<extra_id_7>",
15
+ "<extra_id_8>",
16
+ "<extra_id_9>",
17
+ "<extra_id_10>",
18
+ "<extra_id_11>",
19
+ "<extra_id_12>",
20
+ "<extra_id_13>",
21
+ "<extra_id_14>",
22
+ "<extra_id_15>",
23
+ "<extra_id_16>",
24
+ "<extra_id_17>",
25
+ "<extra_id_18>",
26
+ "<extra_id_19>",
27
+ "<extra_id_20>",
28
+ "<extra_id_21>",
29
+ "<extra_id_22>",
30
+ "<extra_id_23>",
31
+ "<extra_id_24>",
32
+ "<extra_id_25>",
33
+ "<extra_id_26>",
34
+ "<extra_id_27>",
35
+ "<extra_id_28>",
36
+ "<extra_id_29>",
37
+ "<extra_id_30>",
38
+ "<extra_id_31>",
39
+ "<extra_id_32>",
40
+ "<extra_id_33>",
41
+ "<extra_id_34>",
42
+ "<extra_id_35>",
43
+ "<extra_id_36>",
44
+ "<extra_id_37>",
45
+ "<extra_id_38>",
46
+ "<extra_id_39>",
47
+ "<extra_id_40>",
48
+ "<extra_id_41>",
49
+ "<extra_id_42>",
50
+ "<extra_id_43>",
51
+ "<extra_id_44>",
52
+ "<extra_id_45>",
53
+ "<extra_id_46>",
54
+ "<extra_id_47>",
55
+ "<extra_id_48>",
56
+ "<extra_id_49>",
57
+ "<extra_id_50>",
58
+ "<extra_id_51>",
59
+ "<extra_id_52>",
60
+ "<extra_id_53>",
61
+ "<extra_id_54>",
62
+ "<extra_id_55>",
63
+ "<extra_id_56>",
64
+ "<extra_id_57>",
65
+ "<extra_id_58>",
66
+ "<extra_id_59>",
67
+ "<extra_id_60>",
68
+ "<extra_id_61>",
69
+ "<extra_id_62>",
70
+ "<extra_id_63>",
71
+ "<extra_id_64>",
72
+ "<extra_id_65>",
73
+ "<extra_id_66>",
74
+ "<extra_id_67>",
75
+ "<extra_id_68>",
76
+ "<extra_id_69>",
77
+ "<extra_id_70>",
78
+ "<extra_id_71>",
79
+ "<extra_id_72>",
80
+ "<extra_id_73>",
81
+ "<extra_id_74>",
82
+ "<extra_id_75>",
83
+ "<extra_id_76>",
84
+ "<extra_id_77>",
85
+ "<extra_id_78>",
86
+ "<extra_id_79>",
87
+ "<extra_id_80>",
88
+ "<extra_id_81>",
89
+ "<extra_id_82>",
90
+ "<extra_id_83>",
91
+ "<extra_id_84>",
92
+ "<extra_id_85>",
93
+ "<extra_id_86>",
94
+ "<extra_id_87>",
95
+ "<extra_id_88>",
96
+ "<extra_id_89>",
97
+ "<extra_id_90>",
98
+ "<extra_id_91>",
99
+ "<extra_id_92>",
100
+ "<extra_id_93>",
101
+ "<extra_id_94>",
102
+ "<extra_id_95>",
103
+ "<extra_id_96>",
104
+ "<extra_id_97>",
105
+ "<extra_id_98>",
106
+ "<extra_id_99>"
107
+ ],
108
+ "is_local": true,
109
+ "legacy": false,
110
+ "local_files_only": false,
111
+ "model_max_length": 512,
112
+ "pad_token": "<pad>",
113
+ "sp_model_kwargs": {},
114
+ "tokenizer_class": "T5Tokenizer",
115
+ "unk_token": "<unk>"
116
+ }
text_conditioner/config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AnimaTextConditioner",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "min_sequence_length": 512,
5
+ "mlp_ratio": 4.0,
6
+ "model_dim": 1024,
7
+ "num_attention_heads": 16,
8
+ "num_layers": 6,
9
+ "source_dim": 1024,
10
+ "target_dim": 1024,
11
+ "target_vocab_size": 32128,
12
+ "use_layer_norm": false,
13
+ "use_self_attention": true
14
+ }
text_conditioner/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a013e45b20e96c1953496e41cd1fe209a8ec1366da3699c4ff3605b50b9e893f
3
+ size 269339400
text_encoder/config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3Model"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": null,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": null,
10
+ "head_dim": 128,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 1024,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_types": [
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention"
44
+ ],
45
+ "max_position_embeddings": 32768,
46
+ "max_window_layers": 28,
47
+ "model_type": "qwen3",
48
+ "num_attention_heads": 16,
49
+ "num_hidden_layers": 28,
50
+ "num_key_value_heads": 8,
51
+ "pad_token_id": null,
52
+ "rms_norm_eps": 1e-06,
53
+ "rope_parameters": {
54
+ "rope_theta": 1000000.0,
55
+ "rope_type": "default"
56
+ },
57
+ "sliding_window": null,
58
+ "tie_word_embeddings": false,
59
+ "transformers_version": "5.8.0",
60
+ "use_cache": true,
61
+ "use_sliding_window": false,
62
+ "vocab_size": 151936
63
+ }
text_encoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75e8db94adb374b61baca3d898fc2c0d5086d20689a96d4e74576546433a0526
3
+ size 1192133232
tokenizer/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
tokenizer/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:476870a1f2fb6f6a2759a6ede2383bf9d5d738f17844563b65c91965b722ae09
3
+ size 11422924
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": null,
5
+ "clean_up_tokenization_spaces": false,
6
+ "eos_token": "<|im_end|>",
7
+ "errors": "replace",
8
+ "is_local": true,
9
+ "local_files_only": false,
10
+ "model_max_length": 131072,
11
+ "pad_token": "<|endoftext|>",
12
+ "processor_class": "Qwen2_5_VLProcessor",
13
+ "split_special_tokens": false,
14
+ "tokenizer_class": "Qwen2Tokenizer",
15
+ "unk_token": null
16
+ }
transformer/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "CosmosTransformer3DModel",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "adaln_lora_dim": 256,
5
+ "attention_head_dim": 128,
6
+ "concat_padding_mask": true,
7
+ "controlnet_block_every_n": null,
8
+ "crossattn_proj_in_channels": 1024,
9
+ "encoder_hidden_states_channels": 1024,
10
+ "extra_pos_embed_type": null,
11
+ "img_context_dim_in": null,
12
+ "img_context_dim_out": 2048,
13
+ "img_context_num_tokens": 256,
14
+ "in_channels": 16,
15
+ "max_size": [
16
+ 128,
17
+ 240,
18
+ 240
19
+ ],
20
+ "mlp_ratio": 4.0,
21
+ "num_attention_heads": 16,
22
+ "num_layers": 28,
23
+ "out_channels": 16,
24
+ "patch_size": [
25
+ 1,
26
+ 2,
27
+ 2
28
+ ],
29
+ "rope_scale": [
30
+ 1.0,
31
+ 4.0,
32
+ 4.0
33
+ ],
34
+ "text_embed_dim": 1024,
35
+ "use_crossattn_projection": false
36
+ }
transformer/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89f643d148b9a20ddfa34f2ad4015face494aa82ab0b748bfaf6c765778df3ff
3
+ size 3912877104
vae/config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKLQwenImage",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "attn_scales": [],
5
+ "base_dim": 96,
6
+ "dim_mult": [
7
+ 1,
8
+ 2,
9
+ 4,
10
+ 4
11
+ ],
12
+ "dropout": 0.0,
13
+ "input_channels": 3,
14
+ "latents_mean": [
15
+ -0.7571,
16
+ -0.7089,
17
+ -0.9113,
18
+ 0.1075,
19
+ -0.1745,
20
+ 0.9653,
21
+ -0.1517,
22
+ 1.5508,
23
+ 0.4134,
24
+ -0.0715,
25
+ 0.5517,
26
+ -0.3632,
27
+ -0.1922,
28
+ -0.9497,
29
+ 0.2503,
30
+ -0.2921
31
+ ],
32
+ "latents_std": [
33
+ 2.8184,
34
+ 1.4541,
35
+ 2.3275,
36
+ 2.6558,
37
+ 1.2196,
38
+ 1.7708,
39
+ 2.6052,
40
+ 2.0743,
41
+ 3.2687,
42
+ 2.1526,
43
+ 2.8652,
44
+ 1.5579,
45
+ 1.6382,
46
+ 1.1253,
47
+ 2.8251,
48
+ 1.916
49
+ ],
50
+ "num_res_blocks": 2,
51
+ "temperal_downsample": [
52
+ false,
53
+ true,
54
+ true
55
+ ],
56
+ "z_dim": 16
57
+ }
vae/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c8bc8b758c649abef9ea407b95408389a3b2f610d0d10fcb054fe171d0a8344
3
+ size 253806966