bobcastell commited on
Commit
71bb8df
·
verified ·
1 Parent(s): 90a5bf8

Delete 20251103_03-44-07

Browse files
20251103_03-44-07/backup/Dataset_orig/ok (1).mp4 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:537a74fb751a1afc87018a5a12a3d9f27393fb365c3a8924c315cf6bcef8f586
3
- size 623641
 
 
 
 
20251103_03-44-07/backup/Dataset_orig/ok (1).txt DELETED
@@ -1,3 +0,0 @@
1
- The camera pulls back.
2
- She wears a white costume made of bands, garter straps and stockings.
3
- Her costume is bottomless and she has no panties so her vagina is uncovered.
 
 
 
 
20251103_03-44-07/backup/Dataset_orig/ok (10).mp4 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c33395035de62137254af6cb7f263d19ea209e8fb1ac1fbf049bd7de509ecbbc
3
- size 1940953
 
 
 
 
20251103_03-44-07/backup/Dataset_orig/ok (10).txt DELETED
@@ -1,3 +0,0 @@
1
- The camera pulls back.
2
- She wears a policewoman costume with a miniskirt, garter straps and stockings.
3
- Her costume is bottomless and she has no panties so her vagina is uncovered.
 
 
 
 
20251103_03-44-07/backup/Dataset_orig/ok (11).mp4 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9febe4e2f5a3eb3dd13c23d3a56936a77f65733fabb7a3c43d27d032f8d7ac6
3
- size 2245422
 
 
 
 
20251103_03-44-07/backup/Dataset_orig/ok (11).txt DELETED
@@ -1,3 +0,0 @@
1
- The camera pulls back.
2
- She is wearing a bodysuit with garter straps and stockings.
3
- Her bodysuit is bottomless and she has no panties so her vagina is uncovered.
 
 
 
 
20251103_03-44-07/backup/Dataset_orig/ok (12).mp4 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8279223be6cf173228f4d0f0a77b8b8dc202b4210b47fa77912f391babeb1f90
3
- size 1520331
 
 
 
 
20251103_03-44-07/backup/Dataset_orig/ok (12).txt DELETED
@@ -1,3 +0,0 @@
1
- The camera pulls back.
2
- She is wearing a white bodysuit with garter straps and stockings.
3
- Her bodysuit is bottomless and she has no panties so her vagina is uncovered.
 
 
 
 
20251103_03-44-07/backup/Dataset_orig/ok (13).mp4 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:829496968a8ade5cea8f57a468ffd2d1dee9fd4267091908144ac8b0773d442f
3
- size 1354158
 
 
 
 
20251103_03-44-07/backup/dataset_info.txt DELETED
@@ -1,4 +0,0 @@
1
- model_type: wan
2
- dataset_name: Dataset
3
- dataset_path: /root/data/wan/Dataset
4
- backup_time: 2025-11-03 04:55:15
 
 
 
 
 
20251103_03-44-07/epoch25/adapter_config.json DELETED
@@ -1,38 +0,0 @@
1
- {
2
- "alpha_pattern": {},
3
- "auto_mapping": null,
4
- "base_model_name_or_path": null,
5
- "bias": "none",
6
- "eva_config": null,
7
- "exclude_modules": null,
8
- "fan_in_fan_out": false,
9
- "inference_mode": false,
10
- "init_lora_weights": true,
11
- "layer_replication": null,
12
- "layers_pattern": null,
13
- "layers_to_transform": null,
14
- "loftq_config": {},
15
- "lora_alpha": 16,
16
- "lora_bias": false,
17
- "lora_dropout": 0.0,
18
- "megatron_config": null,
19
- "megatron_core": "megatron.core",
20
- "modules_to_save": null,
21
- "peft_type": "LORA",
22
- "r": 16,
23
- "rank_pattern": {},
24
- "revision": null,
25
- "target_modules": [
26
- "v_img",
27
- "k",
28
- "o",
29
- "q",
30
- "v",
31
- "ffn.0",
32
- "k_img",
33
- "ffn.2"
34
- ],
35
- "task_type": null,
36
- "use_dora": false,
37
- "use_rslora": false
38
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20251103_03-44-07/epoch25/adapter_model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:06f0920167318f591e599322d55a4fe3658fe4ecec6b6e8fb40a98967f342cb0
3
- size 179688448
 
 
 
 
20251103_03-44-07/epoch25/wan.toml DELETED
@@ -1,38 +0,0 @@
1
- output_dir = "/root/diffusion-pipe/wan-lora"
2
- dataset = "/root/config/dataset_wan.toml"
3
- epochs = 40
4
- micro_batch_size_per_gpu = 2
5
- pipeline_stages = 1
6
- gradient_accumulation_steps = 4
7
- gradient_clipping = 1.0
8
- warmup_steps = 50
9
- eval_every_n_epochs = 1
10
- eval_before_first_step = true
11
- eval_micro_batch_size_per_gpu = 1
12
- eval_gradient_accumulation_steps = 1
13
- save_every_n_epochs = 5
14
- activation_checkpointing = true
15
- partition_method = "parameters"
16
- save_dtype = "bfloat16"
17
- caching_batch_size = 2
18
- steps_per_print = 1
19
- video_clip_mode = "single_middle"
20
- checkpoint_every_n_epochs = 5
21
-
22
- [model]
23
- type = "wan"
24
- ckpt_path = "/root/diffusion-pipe/ckpts/Wan2.1-I2V-14B-480P"
25
- dtype = "bfloat16"
26
- timestep_sample_method = "logit_normal"
27
-
28
- [adapter]
29
- type = "lora"
30
- rank = 16
31
- dtype = "bfloat16"
32
-
33
- [optimizer]
34
- type = "adamw_optimi"
35
- lr = 0.0001
36
- betas = [ 0.9, 0.99,]
37
- weight_decay = 0.01
38
- eps = 1e-8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20251103_03-44-07/epoch30/adapter_config.json DELETED
@@ -1,38 +0,0 @@
1
- {
2
- "alpha_pattern": {},
3
- "auto_mapping": null,
4
- "base_model_name_or_path": null,
5
- "bias": "none",
6
- "eva_config": null,
7
- "exclude_modules": null,
8
- "fan_in_fan_out": false,
9
- "inference_mode": false,
10
- "init_lora_weights": true,
11
- "layer_replication": null,
12
- "layers_pattern": null,
13
- "layers_to_transform": null,
14
- "loftq_config": {},
15
- "lora_alpha": 16,
16
- "lora_bias": false,
17
- "lora_dropout": 0.0,
18
- "megatron_config": null,
19
- "megatron_core": "megatron.core",
20
- "modules_to_save": null,
21
- "peft_type": "LORA",
22
- "r": 16,
23
- "rank_pattern": {},
24
- "revision": null,
25
- "target_modules": [
26
- "v_img",
27
- "k",
28
- "o",
29
- "q",
30
- "v",
31
- "ffn.0",
32
- "k_img",
33
- "ffn.2"
34
- ],
35
- "task_type": null,
36
- "use_dora": false,
37
- "use_rslora": false
38
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20251103_03-44-07/epoch30/adapter_model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cb64077292945371d447b92bd5ed26fe08c0e4b0df48c57f4ef2bffab8bf126
3
- size 179688448
 
 
 
 
20251103_03-44-07/epoch30/wan.toml DELETED
@@ -1,38 +0,0 @@
1
- output_dir = "/root/diffusion-pipe/wan-lora"
2
- dataset = "/root/config/dataset_wan.toml"
3
- epochs = 40
4
- micro_batch_size_per_gpu = 2
5
- pipeline_stages = 1
6
- gradient_accumulation_steps = 4
7
- gradient_clipping = 1.0
8
- warmup_steps = 50
9
- eval_every_n_epochs = 1
10
- eval_before_first_step = true
11
- eval_micro_batch_size_per_gpu = 1
12
- eval_gradient_accumulation_steps = 1
13
- save_every_n_epochs = 5
14
- activation_checkpointing = true
15
- partition_method = "parameters"
16
- save_dtype = "bfloat16"
17
- caching_batch_size = 2
18
- steps_per_print = 1
19
- video_clip_mode = "single_middle"
20
- checkpoint_every_n_epochs = 5
21
-
22
- [model]
23
- type = "wan"
24
- ckpt_path = "/root/diffusion-pipe/ckpts/Wan2.1-I2V-14B-480P"
25
- dtype = "bfloat16"
26
- timestep_sample_method = "logit_normal"
27
-
28
- [adapter]
29
- type = "lora"
30
- rank = 16
31
- dtype = "bfloat16"
32
-
33
- [optimizer]
34
- type = "adamw_optimi"
35
- lr = 0.0001
36
- betas = [ 0.9, 0.99,]
37
- weight_decay = 0.01
38
- eps = 1e-8