UnifiedHorusRA commited on
Commit
7ca7776
·
verified ·
1 Parent(s): f4cf781

Upload v2.0__hunyuan/README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. v2.0__hunyuan/README.md +152 -0
v2.0__hunyuan/README.md ADDED
@@ -0,0 +1,152 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - art
6
+ ---
7
+
8
+ # Flat Color - Style
9
+
10
+ **Creator**: [motimalu](https://civitai.com/user/motimalu)
11
+ **Type**: LORA
12
+ **Base Model**: Hunyuan Video
13
+ **Version**: v2.0 [hunyuan]
14
+ **Trigger Words**: `flat color, no lineart`
15
+
16
+ **Civitai Model ID**: 1132089
17
+ **Civitai Version ID**: 1315010
18
+
19
+ **Stats (at time of fetch)**:
20
+ * Downloads: 4130
21
+ * Rating: 0 (0 ratings)
22
+ * Favorites: N/A
23
+
24
+ ---
25
+
26
+ ## 📄 Description (Parent Model)
27
+
28
+ Flat Color
29
+ -
30
+ Style
31
+ Trained on images without visible lineart, flat colors, and
32
+ little to no indication of depth.
33
+ ℹ️ LoRA work best when applied to the base models on which they are trained.
34
+ Please read the
35
+ About This Version
36
+ on the appropriate base models
37
+ and workflow/training information.
38
+ This is a small style LoRA I thought would be interesting to try with a v-pred model (noobai v-pred), for the reduced color bleeding and strong blacks in particular.
39
+ The effect is quite nice and easy to evaluate in training, so I've extended the dataset with videos in following versions for text-to-video models like Wan and Hunyuan, and it is what I am generally using to test LoRA training on new models now.
40
+ Recommended prompt structure:
41
+ Positive prompt:
42
+ flat color, no lineart, blending, negative space,
43
+ {{tags}}
44
+ masterpiece, best quality, very aesthetic, newest
45
+
46
+ ## Version Notes (v2.0 [hunyuan])
47
+
48
+ Trained with
49
+ https://github.com/tdrussell/diffusion-pipe
50
+ Training data consists of:
51
+ 42 images as a combination of
52
+ Images used from other versions this model card
53
+ Images extracted as keyframes from several videos
54
+ 19 short video clips ~40 frames each
55
+ Training configs:
56
+ dataset.toml
57
+ # Aspect ratio bucketing settings
58
+ enable_ar_bucket = true
59
+ min_ar = 0.5
60
+ max_ar = 2.0
61
+ num_ar_buckets = 7
62
+
63
+ [[directory]] # IMAGES
64
+ # Path to the directory containing images and their corresponding caption files.
65
+ path = '/mnt/d/huanvideo/training_data/images'
66
+ num_repeats = 5
67
+ resolutions = [1024]
68
+ frame_buckets = [1] # Use 1 frame for images.
69
+
70
+
71
+ [[directory]] # VIDEOS
72
+ # Path to the directory containing videos and their corresponding caption files.
73
+ path = '/mnt/d/huanvideo/training_data/videos'
74
+ num_repeats = 5
75
+ resolutions = [256] # Set video resolution to 256 (e.g., 244p).
76
+ frame_buckets = [33, 49, 81] # Define frame buckets for videos.
77
+ config.toml
78
+ # Dataset config file.
79
+ output_dir = '/mnt/d/huanvideo/training_output'
80
+ dataset = 'dataset.toml'
81
+
82
+ # Training settings
83
+ epochs = 50
84
+ micro_batch_size_per_gpu = 1
85
+ pipeline_stages = 1
86
+ gradient_accumulation_steps = 4
87
+ gradient_clipping = 1.0
88
+ warmup_steps = 100
89
+
90
+ # eval settings
91
+ eval_every_n_epochs = 5
92
+ eval_before_first_step = true
93
+ eval_micro_batch_size_per_gpu = 1
94
+ eval_gradient_accumulation_steps = 1
95
+
96
+ # misc settings
97
+ save_every_n_epochs = 15
98
+ checkpoint_every_n_minutes = 30
99
+ activation_checkpointing = true
100
+ partition_method = 'parameters'
101
+ save_dtype = 'bfloat16'
102
+ caching_batch_size = 1
103
+ steps_per_print = 1
104
+ video_clip_mode = 'single_middle'
105
+
106
+ [model]
107
+ type = 'hunyuan-video'
108
+
109
+ transformer_path = '/mnt/d/huanvideo/models/diffusion_models/hunyuan_video_720_cfgdistill_fp8_e4m3fn.safetensors'
110
+ vae_path = '/mnt/d/huanvideo/models/vae/hunyuan_video_vae_bf16.safetensors'
111
+ llm_path = '/mnt/d/huanvideo/models/llm'
112
+ clip_path = '/mnt/d/huanvideo/models/clip'
113
+
114
+ dtype = 'bfloat16'
115
+ transformer_dtype = 'float8'
116
+ timestep_sample_method = 'logit_normal'
117
+
118
+ [adapter]
119
+ type = 'lora'
120
+ rank = 32
121
+ dtype = 'bfloat16'
122
+
123
+ [optimizer]
124
+ type = 'adamw_optimi'
125
+ lr = 5e-5
126
+ betas = [0.9, 0.99]
127
+ weight_decay = 0.02
128
+ eps = 1e-8
129
+
130
+ ---
131
+
132
+ ## Civitai Links
133
+
134
+ * **[🔗 View This Version on Civitai →](https://civitai.com/models/1132089?modelVersionId=1315010)**
135
+ * [View Full Model Page →](https://civitai.com/models/1132089)
136
+ * [View Creator Profile →](https://civitai.com/user/motimalu)
137
+
138
+ ---
139
+
140
+ ## Preview Images
141
+
142
+ ![Preview Image](./previews/1315010_53170144.mp4)
143
+
144
+ ---
145
+
146
+ ## File Information
147
+
148
+ * **Filename**: `hunyuan_flat_color_v2.safetensors`
149
+ * **Size**: 307.58 MB
150
+ * **Hash (AutoV2)**: `E732F98BE3`
151
+ * **Hash (SHA256)**: `E732F98BE3D6838E3C71A502BCBFDA7852C4B402C2446075D7EA0E0FAC3EEA33`
152
+