Upload folder using huggingface_hub
Browse files- config.json +102 -0
- model.safetensors +3 -0
- optimizer.pt +3 -0
- rng_state_0.pth +3 -0
- rng_state_1.pth +3 -0
- rng_state_2.pth +3 -0
- rng_state_3.pth +3 -0
- rng_state_4.pth +3 -0
- rng_state_5.pth +3 -0
- rng_state_6.pth +3 -0
- rng_state_7.pth +3 -0
- scheduler.pt +3 -0
- trainer_state.json +0 -0
- training_args.bin +3 -0
config.json
ADDED
|
@@ -0,0 +1,102 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_gradient_checkpointing": true,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"MetaQuery"
|
| 5 |
+
],
|
| 6 |
+
"attn_implementation": "sdpa",
|
| 7 |
+
"audio_uncond_ratio": 0.05,
|
| 8 |
+
"audio_window_size": 2,
|
| 9 |
+
"connector_num_hidden_layers": 8,
|
| 10 |
+
"context_uncond_ratio": 0.2,
|
| 11 |
+
"decode_noise_scale": 0.025,
|
| 12 |
+
"decode_timestep": 0.05,
|
| 13 |
+
"diffusion_forcing": {
|
| 14 |
+
"block_causal": false,
|
| 15 |
+
"enable": true,
|
| 16 |
+
"few_steps": false,
|
| 17 |
+
"is_neg_posi_share_kv": true,
|
| 18 |
+
"kv_save_step": -1,
|
| 19 |
+
"num_lookback_chunks": 2,
|
| 20 |
+
"prefix_timestep": -1,
|
| 21 |
+
"ref_uncond_ratio": 0.1,
|
| 22 |
+
"temp_chunk": 3,
|
| 23 |
+
"use_ref_as_sink": true
|
| 24 |
+
},
|
| 25 |
+
"diffusion_model_id": "Lightricks/LTX-Video",
|
| 26 |
+
"diffusion_model_path": "./pretrained_models/Wan2.1-T2V-14B/",
|
| 27 |
+
"gen_fps": 25.0,
|
| 28 |
+
"in_channels": 16,
|
| 29 |
+
"infer": false,
|
| 30 |
+
"joint_uncond_ratio": 0.1,
|
| 31 |
+
"latent_spatial_size": 56,
|
| 32 |
+
"latent_temporal_size": 16,
|
| 33 |
+
"local_files_only": true,
|
| 34 |
+
"lora_dict": {
|
| 35 |
+
"adapter_name": "context",
|
| 36 |
+
"enable": true,
|
| 37 |
+
"network_alpha": 128,
|
| 38 |
+
"rank": 128,
|
| 39 |
+
"target_modules": [
|
| 40 |
+
"q",
|
| 41 |
+
"k",
|
| 42 |
+
"v",
|
| 43 |
+
"o",
|
| 44 |
+
"ffn.0",
|
| 45 |
+
"ffn.2"
|
| 46 |
+
]
|
| 47 |
+
},
|
| 48 |
+
"loss_type": "flow",
|
| 49 |
+
"max_input_text_tokens": 256,
|
| 50 |
+
"mllm_id": "./pretrained_models/Qwen2.5-Omni-7B",
|
| 51 |
+
"mllm_local_path": null,
|
| 52 |
+
"model_type": "metaquery",
|
| 53 |
+
"modules_to_freeze": [
|
| 54 |
+
"audio_encoder",
|
| 55 |
+
"pipe.vae",
|
| 56 |
+
"pipe.dit",
|
| 57 |
+
"pipe.text_encoder",
|
| 58 |
+
"mllm.mllm_backbone"
|
| 59 |
+
],
|
| 60 |
+
"modules_to_unfreeze": [
|
| 61 |
+
"mllm.connector",
|
| 62 |
+
"mllm.mllm_backbone.model.embed_tokens"
|
| 63 |
+
],
|
| 64 |
+
"negative_text_prompt": "Vivid color tones, background/camera moving quickly, screen switching, subtitles and special effects, mutation, overexposed, static, blurred details, abrupt and extreme head shift, subtitles, style, work, painting, image, still, overall grayish, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn face, deformed, disfigured, malformed limbs, fingers merging, motionless image, chaotic background, three legs, crowded background with many people, walking backward",
|
| 65 |
+
"noise_scheduler_id": "Lightricks/LTX-Video",
|
| 66 |
+
"noise_to_first_frame": 0.25,
|
| 67 |
+
"num_metaqueries": 128,
|
| 68 |
+
"omniavatar_dict": {
|
| 69 |
+
"init_lora_weights": "kaiming",
|
| 70 |
+
"lora_alpha": 64,
|
| 71 |
+
"lora_rank": 128,
|
| 72 |
+
"lora_reinit": false,
|
| 73 |
+
"lora_target_modules": "q,k,v,o,ffn.0,ffn.2",
|
| 74 |
+
"merged": true,
|
| 75 |
+
"weights_path": "./pretrained_models/OmniAvatar-14B/pytorch_model.pt"
|
| 76 |
+
},
|
| 77 |
+
"prev_temp_chunk": 4,
|
| 78 |
+
"retify_to_last_frame": true,
|
| 79 |
+
"scheduler_id": "Lightricks/LTX-Video",
|
| 80 |
+
"scheduler_type": "euler",
|
| 81 |
+
"system_prompt": "You are a specialized system for generating descriptions of an agent's non-verbal behavior during face-to-face conversations. Your output will be used as learnable textual control tokens for a video generation model.\nInput Format: You will receive two historical context segments: [First] Agent's historical context (audio always available; video often unavailable): - Recent audio of the agent's speech or vocalization - Recent agent video frames if available (in most cases not present) [Second] User's historical context (audio always available; video may be available): - Recent audio of the user's speech - Recent user video frames when provided\n[Output] - Produce non-verbal behavior description encoded as learnable textual cues - This description should predict the agent's upcoming head pose tendencies, micro-movements, gaze patterns, and facial dynamics\nKey Principles: - React naturally to the conversational dynamics captured in [Second] (user's timing, tone, emotional state, visual cues if available) - Focus solely on the agent's visible head region: head orientation, gaze direction, facial muscle movements - Do not assume any specific upcoming speech timing, duration, or transitions, as future audio is unknown\nOutput Structure (Text after each colon explains how to answer that item): 1. State: Indicate whether the agent is more likely to be in a \"Speaking\", \"Listening\" or \"Turn-switching\"(rapid shifts expected) state based on the historical conversational flow. If unclear, choose the state suggested by the most recent history. 2. Overall demeanor: Briefly characterize the intended non-verbal attitude (e.g., engaged, thoughtful, friendly, contemplative, etc.) 3. Arousal level: From low to high, which determines the intensity, speed, and amplitude of the agent's movements described below 4. Head motion: Describe the agent's upcoming natural head movements (posture maintenance, nods, tilts, rhythmic adjustments, etc.) 5. Gaze: Outline expected gaze behavior, specifying focus targets (e.g., direct eye contact, averting gaze for thought, scanning), referencing to the user\u2019s location if their video is present in [Second]. 6. Facial expression: Describe the agent\u2019s facial expression tendencies, covering the likely baseline (neutral, expressive, or relaxed, etc.) and potential changes in the eyes, mouth, or eyebrows that correspond to the tone and intensity of the conversation. 7. Listening / Speaking Behavior: - If listening: give a more detailed description of active listening behavior with above aspects - If speaking: give a concise description of speech-related coordination with above aspects\nConstraints: - Do NOT transcribe or summarize speech content - Do NOT describe the user's behavior, appearance and background - Do NOT describe the agent's appearance, background - Focus on the upcoming response behavior in detail",
|
| 82 |
+
"text_encoder_id": "./pretrained_models/Wan2.1-T2V-14B",
|
| 83 |
+
"text_encoder_local_path": null,
|
| 84 |
+
"text_encoder_path": "./pretrained_models/Wan2.1-T2V-14B/models_t5_umt5-xxl-enc-bf16.pth",
|
| 85 |
+
"text_prompt": "A realistic video of a person communicating with another person by front-facing camera, with dynamic facial expression and rhythmichead motion that complement his talking or listening responses.",
|
| 86 |
+
"tokenizer_local_path": null,
|
| 87 |
+
"torch_dtype": "bfloat16",
|
| 88 |
+
"transformers_version": "4.52.3",
|
| 89 |
+
"use_context": true,
|
| 90 |
+
"use_region_loss": {
|
| 91 |
+
"enable": false,
|
| 92 |
+
"eye": 0.3,
|
| 93 |
+
"head": 0.2,
|
| 94 |
+
"mouth": 0.5
|
| 95 |
+
},
|
| 96 |
+
"use_target_audio": true,
|
| 97 |
+
"vae_downsample_f": 8,
|
| 98 |
+
"vae_downsample_t": 4,
|
| 99 |
+
"vae_id": "Lightricks/LTX-Video",
|
| 100 |
+
"vae_path": "./pretrained_models/Wan2.1-T2V-14B/Wan2.1_VAE.pth",
|
| 101 |
+
"wav2vec_path": "./pretrained_models/wav2vec2-base-960h"
|
| 102 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7a86df145ad681d84eed00375f1baf76a9b711952e7fc1fac5cb9a17bfe88e76
|
| 3 |
+
size 3465287048
|
optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:57f56778d58f255346e1428bd71b9374f94fa83e54aa28d70d21a51ce1a7a24a
|
| 3 |
+
size 6931161524
|
rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b371c76d3a12ca423aa5dbbd3ec2bafc19de4c20c4b9dd06285d6127e8d8638
|
| 3 |
+
size 16389
|
rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:43f854c9fb767921e25aa1b3894ff165baadee2166d4752d758febc790909d3d
|
| 3 |
+
size 16389
|
rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b7736fb612d59ef87a0e9a60fc7ededddf27ce4fbe0c8a7b8af022bf418f604a
|
| 3 |
+
size 16389
|
rng_state_3.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:030c48620b1d8864fa197defe20550b200efb42aab9374eefed3135fdd25f919
|
| 3 |
+
size 16389
|
rng_state_4.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bd9013cebcf5b9095fe8a54eb281f126d567cfaf4037f3bfb2211da5e59b1faf
|
| 3 |
+
size 16389
|
rng_state_5.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b156a90ea8d3274fde8c6028e25d18e1098699e4e5115e347368a408c2c09b3a
|
| 3 |
+
size 16389
|
rng_state_6.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b60ca1e0da875a37e38cc28af99674602c0c9db8a6a4202817561ef59ce65af7
|
| 3 |
+
size 16389
|
rng_state_7.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4240bfb48d6759b1e4c393dc4f26cccb1e6e8619b6e8f8b6e9fb4078208b8491
|
| 3 |
+
size 16389
|
scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d10983ea22496cb85733e8a649c58f9a63decf9d05d1b318b06136c3e64e90a5
|
| 3 |
+
size 1465
|
trainer_state.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:955f7d7bfb861f692ca588ef2ab045c1e67871fb4f63c4dea96395365a567bbc
|
| 3 |
+
size 6225
|