Alissonerdx commited on
Commit
3b2424e
·
verified ·
1 Parent(s): 002ae79

Upload folder using huggingface_hub

Browse files
prodigy_v2_128_pure_face/samples/step_000000_1.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e76032d2918e690165e37f482a1dd35faddca87df68f3de637a333b45196c6c7
3
- size 518548
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f743287dc46c7571dd193959ebfa8bc8a89aae6b153bf2d534603c7eaaa68ee6
3
+ size 542950
prodigy_v2_128_pure_face/samples/step_000100_1.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1180e5a59b104a70705518ac0703efdbffa85f5b8e2e33f8c8f6ea243e27bded
3
- size 523447
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:986cdc2041a72ea45286cb987936b6d3e12707ab0dc98c09f63ccc9acf44dabe
3
+ size 631164
prodigy_v2_128_pure_face/samples/step_000200_1.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c3ede157706c4bf36541f396f7255f40677f1467c5929f856d578c0fc698c89
3
- size 523878
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9070d45dcb1769ea118aa55c0db099db673cc196556eb2fb92831bcb330a629
3
+ size 596063
prodigy_v2_128_pure_face/samples/step_000300_1.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6a10bd32fb6b761522f38648bd26885ea981cffb9753f3545d10f5dc62424ff
3
- size 510038
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:260e346a5087fffc110a72ff0ac8e1f4eba4e4ea7bf5205ab9954ea99298e5cf
3
+ size 554016
prodigy_v2_128_pure_face/training_config.yaml CHANGED
@@ -10,7 +10,7 @@ data:
10
  num_dataloader_workers: 2
11
  preprocessed_data_root: .precomputed
12
  flow_matching:
13
- timestep_sampling_mode: power_law
14
  timestep_sampling_params: {}
15
  hub:
16
  hub_model_id: null
@@ -32,22 +32,22 @@ lora:
32
  - ff.net.2
33
  - adaln_single.linear
34
  model:
35
- load_checkpoint: /workspace/LTX-2/packages/ltx-trainer/outputs/ltx2_v2v_ic_lora_head_swap_v2_prodigy_128/checkpoints
36
  model_path: /workspace/models/ltx-2-19b-dev.safetensors
37
  text_encoder_path: /workspace/models/gemma
38
  training_mode: lora
39
  optimization:
40
  batch_size: 1
41
  enable_gradient_checkpointing: true
42
- gradient_accumulation_steps: 4
43
  learning_rate: 1.0
44
  max_grad_norm: 1.0
45
  optimizer_params: {}
46
  optimizer_type: prodigy
47
  scheduler_params: {}
48
- scheduler_type: linear
49
- steps: 2000
50
- output_dir: /workspace/LTX-2/packages/ltx-trainer/outputs/ltx2_v2v_ic_lora_head_swap_v2_prodigy_128_power_law
51
  seed: 42
52
  training_strategy:
53
  first_frame_conditioning_p: 1.0
 
10
  num_dataloader_workers: 2
11
  preprocessed_data_root: .precomputed
12
  flow_matching:
13
+ timestep_sampling_mode: shifted_logit_normal
14
  timestep_sampling_params: {}
15
  hub:
16
  hub_model_id: null
 
32
  - ff.net.2
33
  - adaln_single.linear
34
  model:
35
+ load_checkpoint: null
36
  model_path: /workspace/models/ltx-2-19b-dev.safetensors
37
  text_encoder_path: /workspace/models/gemma
38
  training_mode: lora
39
  optimization:
40
  batch_size: 1
41
  enable_gradient_checkpointing: true
42
+ gradient_accumulation_steps: 1
43
  learning_rate: 1.0
44
  max_grad_norm: 1.0
45
  optimizer_params: {}
46
  optimizer_type: prodigy
47
  scheduler_params: {}
48
+ scheduler_type: constant
49
+ steps: 15000
50
+ output_dir: /workspace/LTX-2/packages/ltx-trainer/outputs/ltx2_v2v_ic_lora_head_swap_v2_prodigy_128
51
  seed: 42
52
  training_strategy:
53
  first_frame_conditioning_p: 1.0