Upload folder using huggingface_hub
Browse files- .gitattributes +2 -0
- adamw8/checkpoints/lora_weights_step_00250.safetensors +1 -1
- adamw8/checkpoints/lora_weights_step_00500.safetensors +1 -1
- adamw8/checkpoints/lora_weights_step_02500.safetensors +3 -0
- adamw8/samples/step_000250_1.mp4 +2 -2
- adamw8/samples/step_000250_2.mp4 +2 -2
- adamw8/samples/step_000500_1.mp4 +2 -2
- adamw8/samples/step_000500_2.mp4 +2 -2
- adamw8/samples/step_002500_1.mp4 +3 -0
- adamw8/samples/step_002500_2.mp4 +3 -0
- adamw8/training_config.yaml +7 -10
.gitattributes
CHANGED
|
@@ -501,3 +501,5 @@ adamw8/samples/step_002000_1.mp4 filter=lfs diff=lfs merge=lfs -text
|
|
| 501 |
adamw8/samples/step_002000_2.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 502 |
adamw8/samples/step_002250_1.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 503 |
adamw8/samples/step_002250_2.mp4 filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 501 |
adamw8/samples/step_002000_2.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 502 |
adamw8/samples/step_002250_1.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 503 |
adamw8/samples/step_002250_2.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 504 |
+
adamw8/samples/step_002500_1.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 505 |
+
adamw8/samples/step_002500_2.mp4 filter=lfs diff=lfs merge=lfs -text
|
adamw8/checkpoints/lora_weights_step_00250.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1316096704
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2e8aefc0288359a1f6f7959657dcce88ab999b30941ee0f8e35bab8f35563b43
|
| 3 |
size 1316096704
|
adamw8/checkpoints/lora_weights_step_00500.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1316096704
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8cc583094874ee2366eb01eb057e3477acf017edd7bf21036bc779bebaa6b7b5
|
| 3 |
size 1316096704
|
adamw8/checkpoints/lora_weights_step_02500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1395e51f888f13384d62bf420fba5de25772e2b8f0b5c22809d65ad50bc2ffcd
|
| 3 |
+
size 1316096704
|
adamw8/samples/step_000250_1.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4779fc2e7a45a0c2383c7fee6c5b123d9abe3fed0b0f645aa754e6e2fdb2fa4a
|
| 3 |
+
size 836703
|
adamw8/samples/step_000250_2.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28c982f601c3ac689f8a2a83fa5b492e0a0b4387b8e77398495aa50af4d9a05d
|
| 3 |
+
size 964086
|
adamw8/samples/step_000500_1.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:01dea578771eb60d2649213d0b2f7d9d4d6b0c9c115d52adef77b68a374c9b24
|
| 3 |
+
size 824876
|
adamw8/samples/step_000500_2.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f38b9cb8467a274f2d0e9c17340bff562647e0ffba65e374874fd11fc4fb8900
|
| 3 |
+
size 944663
|
adamw8/samples/step_002500_1.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0fe3694b5331c85a8b32fc099994497a00c76489166ecdedb50e1712ac226053
|
| 3 |
+
size 855783
|
adamw8/samples/step_002500_2.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cab863526c0c54b88ec82fb4e96f9e887571fcaeb6afbfef959ce88f6387db15
|
| 3 |
+
size 962301
|
adamw8/training_config.yaml
CHANGED
|
@@ -10,8 +10,10 @@ data:
|
|
| 10 |
num_dataloader_workers: 2
|
| 11 |
preprocessed_data_root: .precomputed
|
| 12 |
flow_matching:
|
| 13 |
-
timestep_sampling_mode:
|
| 14 |
-
timestep_sampling_params:
|
|
|
|
|
|
|
| 15 |
hub:
|
| 16 |
hub_model_id: Alissonerdx/BFS-LTX2
|
| 17 |
push_to_hub: true
|
|
@@ -31,13 +33,8 @@ lora:
|
|
| 31 |
- ff.net.0.proj
|
| 32 |
- ff.net.2
|
| 33 |
- adaln_single.linear
|
| 34 |
-
- video_embeddings_connector.transformer_1d_blocks.0.attn1.to_q
|
| 35 |
-
- video_embeddings_connector.transformer_1d_blocks.0.attn1.to_k
|
| 36 |
-
- video_embeddings_connector.transformer_1d_blocks.1.attn1.to_q
|
| 37 |
-
- video_embeddings_connector.transformer_1d_blocks.1.attn1.to_k
|
| 38 |
-
- scale_shift_table
|
| 39 |
model:
|
| 40 |
-
load_checkpoint: /train/LTX-2/packages/ltx-trainer/outputs/
|
| 41 |
model_path: /train/models/ltx-2-19b-dev.safetensors
|
| 42 |
text_encoder_path: /train/models/gemma
|
| 43 |
training_mode: lora
|
|
@@ -45,11 +42,11 @@ optimization:
|
|
| 45 |
batch_size: 1
|
| 46 |
enable_gradient_checkpointing: true
|
| 47 |
gradient_accumulation_steps: 1
|
| 48 |
-
learning_rate: 0.
|
| 49 |
max_grad_norm: 1.0
|
| 50 |
optimizer_type: adamw8bit
|
| 51 |
scheduler_params: {}
|
| 52 |
-
scheduler_type:
|
| 53 |
steps: 3000
|
| 54 |
output_dir: /train/LTX-2/packages/ltx-trainer/outputs/ltx2_v2v_ic_lora_head_swap_128_adamw8_5
|
| 55 |
seed: 42
|
|
|
|
| 10 |
num_dataloader_workers: 2
|
| 11 |
preprocessed_data_root: .precomputed
|
| 12 |
flow_matching:
|
| 13 |
+
timestep_sampling_mode: uniform
|
| 14 |
+
timestep_sampling_params:
|
| 15 |
+
max_value: 0.8
|
| 16 |
+
min_value: 0.1
|
| 17 |
hub:
|
| 18 |
hub_model_id: Alissonerdx/BFS-LTX2
|
| 19 |
push_to_hub: true
|
|
|
|
| 33 |
- ff.net.0.proj
|
| 34 |
- ff.net.2
|
| 35 |
- adaln_single.linear
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
model:
|
| 37 |
+
load_checkpoint: /train/LTX-2/packages/ltx-trainer/outputs/ltx2_v2v_ic_lora_head_swap_128_adamw8_5/checkpoints
|
| 38 |
model_path: /train/models/ltx-2-19b-dev.safetensors
|
| 39 |
text_encoder_path: /train/models/gemma
|
| 40 |
training_mode: lora
|
|
|
|
| 42 |
batch_size: 1
|
| 43 |
enable_gradient_checkpointing: true
|
| 44 |
gradient_accumulation_steps: 1
|
| 45 |
+
learning_rate: 0.0001
|
| 46 |
max_grad_norm: 1.0
|
| 47 |
optimizer_type: adamw8bit
|
| 48 |
scheduler_params: {}
|
| 49 |
+
scheduler_type: linear
|
| 50 |
steps: 3000
|
| 51 |
output_dir: /train/LTX-2/packages/ltx-trainer/outputs/ltx2_v2v_ic_lora_head_swap_128_adamw8_5
|
| 52 |
seed: 42
|