UnifiedHorusRA commited on
Commit
d905a83
·
verified ·
1 Parent(s): 94fdb03

Upload v2.0_wan-t2v-1.3b/README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. v2.0_wan-t2v-1.3b/README.md +154 -0
v2.0_wan-t2v-1.3b/README.md ADDED
@@ -0,0 +1,154 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - art
6
+ ---
7
+
8
+ # Flat Color - Style
9
+
10
+ **Creator**: [motimalu](https://civitai.com/user/motimalu)
11
+ **Type**: LORA
12
+ **Base Model**: Wan Video 1.3B t2v
13
+ **Version**: v2.0 [wan-t2v-1.3b]
14
+ **Trigger Words**: `flat color, no lineart`
15
+
16
+ **Civitai Model ID**: 1132089
17
+ **Civitai Version ID**: 1525407
18
+
19
+ **Stats (at time of fetch for this version)**:
20
+ * Downloads: 917
21
+ * Rating: 0 (0 ratings)
22
+ * Favorites: N/A
23
+
24
+ ---
25
+
26
+ ## 📄 Description (Parent Model)
27
+
28
+ Flat Color
29
+ -
30
+ Style
31
+ Trained on images without visible lineart, flat colors, and
32
+ little to no indication of depth.
33
+ ℹ️ LoRA work best when applied to the base models on which they are trained.
34
+ Please read the
35
+ About This Version
36
+ on the appropriate base models
37
+ and workflow/training information.
38
+ This is a small style LoRA I thought would be interesting to try with a v-pred model (noobai v-pred), for the reduced color bleeding and strong blacks in particular.
39
+ The effect is quite nice and easy to evaluate in training, so I've extended the dataset with videos in following versions for text-to-video models like Wan and Hunyuan, and it is what I am generally using to test LoRA training on new models now.
40
+ Recommended prompt structure:
41
+ Positive prompt:
42
+ flat color, no lineart, blending, negative space,
43
+ {{tags}}
44
+ masterpiece, best quality, very aesthetic, newest
45
+
46
+ ## Version Notes (v2.0 [wan-t2v-1.3b])
47
+
48
+ [WAN 1.3B] LoRA
49
+ Trained with
50
+ diffusion-pipe
51
+ on
52
+ Wan2.1-T2V-1.3B
53
+ Increased video training resolution to 512
54
+ Lowered video FPS to 16
55
+ Updated frame_buckets to match video frame counts
56
+ Seems usable for both text-to-video and image-to-video workflows with Wan
57
+ Text to Video previews generated with
58
+ ComfyUI_examples/wan/#text-to-video
59
+ Loading the LoRA with LoraLoaderModelOnly node and using the fp16 1.3B
60
+ wan2.1_t2v_1.3B_fp16.safetensors
61
+ Image to Video previews generated with
62
+ ComfyUI_examples/wan/#image-to-video
63
+ Using image generations from the IL/Noobai versions of this model card
64
+ dataset.toml
65
+ # Resolution settings.
66
+ resolutions = [512]
67
+
68
+ # Aspect ratio bucketing settings
69
+ enable_ar_bucket = true
70
+ min_ar = 0.5
71
+ max_ar = 2.0
72
+ num_ar_buckets = 7
73
+
74
+ # Frame buckets (1 is for images)
75
+ frame_buckets = [1]
76
+
77
+ [[directory]] # IMAGES
78
+ # Path to the directory containing images and their corresponding caption files.
79
+ path = '/mnt/d/huanvideo/training_data/images'
80
+ num_repeats = 5
81
+ resolutions = [720]
82
+ frame_buckets = [1] # Use 1 frame for images.
83
+
84
+ [[directory]] # VIDEOS
85
+ # Path to the directory containing videos and their corresponding caption files.
86
+ path = '/mnt/d/huanvideo/training_data/videos'
87
+ num_repeats = 5
88
+ resolutions = [512] # Set video resolution
89
+ frame_buckets = [28, 31, 32, 36, 42, 43, 48, 50, 53]
90
+ config.toml
91
+ # Dataset config file.
92
+ output_dir = '/mnt/d/wan/training_output'
93
+ dataset = 'dataset.toml'
94
+
95
+ # Training settings
96
+ epochs = 50
97
+ micro_batch_size_per_gpu = 1
98
+ pipeline_stages = 1
99
+ gradient_accumulation_steps = 4
100
+ gradient_clipping = 1.0
101
+ warmup_steps = 100
102
+
103
+ # eval settings
104
+ eval_every_n_epochs = 5
105
+ eval_before_first_step = true
106
+ eval_micro_batch_size_per_gpu = 1
107
+ eval_gradient_accumulation_steps = 1
108
+
109
+ # misc settings
110
+ save_every_n_epochs = 5
111
+ checkpoint_every_n_minutes = 30
112
+ activation_checkpointing = true
113
+ partition_method = 'parameters'
114
+ save_dtype = 'bfloat16'
115
+ caching_batch_size = 1
116
+ steps_per_print = 1
117
+ video_clip_mode = 'single_middle'
118
+
119
+ [model]
120
+ type = 'wan'
121
+ ckpt_path = '../Wan2.1-T2V-1.3B'
122
+ dtype = 'bfloat16'
123
+ # You can use fp8 for the transformer when training LoRA.
124
+ transformer_dtype = 'float8'
125
+ timestep_sample_method = 'logit_normal'
126
+
127
+ [adapter]
128
+ type = 'lora'
129
+ rank = 32
130
+ dtype = 'bfloat16'
131
+
132
+ [optimizer]
133
+ type = 'adamw_optimi'
134
+ lr = 5e-5
135
+ betas = [0.9, 0.99]
136
+ weight_decay = 0.02
137
+ eps = 1e-8
138
+
139
+ ---
140
+
141
+ ## Civitai Links
142
+
143
+ * **[🔗 View This Version on Civitai →](https://civitai.com/models/1132089?modelVersionId=1525407)**
144
+ * [View Full Model Page →](https://civitai.com/models/1132089)
145
+ * [View Creator Profile →](https://civitai.com/user/motimalu)
146
+
147
+ ---
148
+
149
+ ## File Information
150
+
151
+ * **Filename**: `wan_flat_color_1.3b_v2.safetensors`
152
+ * **Size**: 83.51 MB
153
+ * **Hash (AutoV2)**: `A95BBBFB76`
154
+ * **Hash (SHA256)**: `A95BBBFB764802492702ABE98D4841415517708477520CDFACF6A2D81A87357D`