Upload folder using huggingface_hub
Browse files
prodigy_v2_128_pure_face/samples/step_000000_1.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f743287dc46c7571dd193959ebfa8bc8a89aae6b153bf2d534603c7eaaa68ee6
|
| 3 |
+
size 542950
|
prodigy_v2_128_pure_face/samples/step_000100_1.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:986cdc2041a72ea45286cb987936b6d3e12707ab0dc98c09f63ccc9acf44dabe
|
| 3 |
+
size 631164
|
prodigy_v2_128_pure_face/samples/step_000200_1.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c9070d45dcb1769ea118aa55c0db099db673cc196556eb2fb92831bcb330a629
|
| 3 |
+
size 596063
|
prodigy_v2_128_pure_face/samples/step_000300_1.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:260e346a5087fffc110a72ff0ac8e1f4eba4e4ea7bf5205ab9954ea99298e5cf
|
| 3 |
+
size 554016
|
prodigy_v2_128_pure_face/training_config.yaml
CHANGED
|
@@ -10,7 +10,7 @@ data:
|
|
| 10 |
num_dataloader_workers: 2
|
| 11 |
preprocessed_data_root: .precomputed
|
| 12 |
flow_matching:
|
| 13 |
-
timestep_sampling_mode:
|
| 14 |
timestep_sampling_params: {}
|
| 15 |
hub:
|
| 16 |
hub_model_id: null
|
|
@@ -32,22 +32,22 @@ lora:
|
|
| 32 |
- ff.net.2
|
| 33 |
- adaln_single.linear
|
| 34 |
model:
|
| 35 |
-
load_checkpoint:
|
| 36 |
model_path: /workspace/models/ltx-2-19b-dev.safetensors
|
| 37 |
text_encoder_path: /workspace/models/gemma
|
| 38 |
training_mode: lora
|
| 39 |
optimization:
|
| 40 |
batch_size: 1
|
| 41 |
enable_gradient_checkpointing: true
|
| 42 |
-
gradient_accumulation_steps:
|
| 43 |
learning_rate: 1.0
|
| 44 |
max_grad_norm: 1.0
|
| 45 |
optimizer_params: {}
|
| 46 |
optimizer_type: prodigy
|
| 47 |
scheduler_params: {}
|
| 48 |
-
scheduler_type:
|
| 49 |
-
steps:
|
| 50 |
-
output_dir: /workspace/LTX-2/packages/ltx-trainer/outputs/
|
| 51 |
seed: 42
|
| 52 |
training_strategy:
|
| 53 |
first_frame_conditioning_p: 1.0
|
|
|
|
| 10 |
num_dataloader_workers: 2
|
| 11 |
preprocessed_data_root: .precomputed
|
| 12 |
flow_matching:
|
| 13 |
+
timestep_sampling_mode: shifted_logit_normal
|
| 14 |
timestep_sampling_params: {}
|
| 15 |
hub:
|
| 16 |
hub_model_id: null
|
|
|
|
| 32 |
- ff.net.2
|
| 33 |
- adaln_single.linear
|
| 34 |
model:
|
| 35 |
+
load_checkpoint: null
|
| 36 |
model_path: /workspace/models/ltx-2-19b-dev.safetensors
|
| 37 |
text_encoder_path: /workspace/models/gemma
|
| 38 |
training_mode: lora
|
| 39 |
optimization:
|
| 40 |
batch_size: 1
|
| 41 |
enable_gradient_checkpointing: true
|
| 42 |
+
gradient_accumulation_steps: 1
|
| 43 |
learning_rate: 1.0
|
| 44 |
max_grad_norm: 1.0
|
| 45 |
optimizer_params: {}
|
| 46 |
optimizer_type: prodigy
|
| 47 |
scheduler_params: {}
|
| 48 |
+
scheduler_type: constant
|
| 49 |
+
steps: 15000
|
| 50 |
+
output_dir: /workspace/LTX-2/packages/ltx-trainer/outputs/ltx2_v2v_ic_lora_head_swap_v2_prodigy_128
|
| 51 |
seed: 42
|
| 52 |
training_strategy:
|
| 53 |
first_frame_conditioning_p: 1.0
|