UnifiedHorusRA commited on
Commit
af4e5dd
ยท
verified ยท
1 Parent(s): 803380f

Upload LTXV_13b_0.9.7_I2V/README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. LTXV_13b_0.9.7_I2V/README.md +275 -0
LTXV_13b_0.9.7_I2V/README.md ADDED
@@ -0,0 +1,275 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - art
6
+ ---
7
+
8
+ # Live Wallpaper Style
9
+
10
+ **Creator**: [NRDX](https://civitai.com/user/NRDX)
11
+ **Type**: LORA
12
+ **Base Model**: LTXV
13
+ **Version**: LTXV 13b 0.9.7 I2V
14
+ **Trigger Words**: `lvwpr`
15
+
16
+ **Civitai Model ID**: 1264662
17
+ **Civitai Version ID**: 1853133
18
+
19
+ **Stats (at time of fetch)**:
20
+ * Downloads: 273
21
+ * Rating: 0 (0 ratings)
22
+ * Favorites: N/A
23
+
24
+ ---
25
+
26
+ ## ๐Ÿ“„ Description (Parent Model)
27
+
28
+ The goal of this lora is to reproduce the video style similar to live wallpaper, for those who play league of legends remember the launcher opening videos, that's the goal, but you can also use it to create your lofi videos :D enjoy.
29
+ [Wan2.2 TI2V 5B - Motion Optimized Edition]
30
+ Trained on 51 curated videos (24fps, 96 frames) for 5,000 steps across 100 epochs with rank 48. Optimized specifically for Wan2.2's unified TI2V 5B dense model and high-compression VAE.
31
+ My Workflow (It's not organized, the important thing is that it works hahaha):
32
+ ๐ŸŽฎ Live Wallpaper LoRA - Wan2.2 5B (Workflow) | Patreon
33
+ Loop Workflow:
34
+ WAN 2.2 5b WhiteRabbit InterpLoop - v1.0 - Hardline | Wan Video Workflows | Civitai
35
+ Trigger word:
36
+ l1v3w4llp4p3r
37
+ [Wan2.2 I2V A14B - Full Timestep Edition]
38
+ Trained on 301 curated videos (256px, 16fps, 49 frames) for 24 hours using Diffusion Pipe with Automagic optimizer, rank 64. Uses extended timestep range (0-1) instead of standard (0-0.875), enabling compatibility with both Low and High models despite training only on Low model.
39
+ Trigger word:
40
+ l1v3w4llp4p3r
41
+ Works excellently with LightX2V v2 (256 rank) for faster inference - recommended starting strength: 2.0 for both LoRAs to avoid artifacts. Loop workflows not yet tested.
42
+ [Wan I2V 720P Fast Fusion - 4 (or more) steps]
43
+ Wan I2V 720P Fast Fusion combines 2 Live Wallpaper LoRA (1 Exclusive) with Lightx2v, AccVid, MoviiGen and Pusa LoRAs for ultra-fast 4+ steps generation while maintaining cinematic quality.
44
+ ๐Ÿš€
45
+ Lightx2v LoRA
46
+ โ€“ accelerates generation by 20x through 4-step distillation, enabling sub 2-minute videos on RTX 4090 with only 8GB VRAM requirements.
47
+ ๐ŸŽฌ
48
+ AccVid LoRA
49
+ โ€“ improves motion accuracy and dynamics for expressive sequences.
50
+ ๐ŸŒŒ
51
+ MoviiGen LoRA
52
+ โ€“ adds cinematic depth and flow to animation, enhancing visual storytelling.
53
+ ๐Ÿง 
54
+ Pusa LoRA
55
+ โ€“ provides fine-grained temporal control with zero-shot multi-task capabilities (start-end frames, video extension) while achieving 87.32% VBench score.
56
+ ๐Ÿง 
57
+ Wan I2V 720p (14B)
58
+ base model โ€“ providing strong temporal consistency and high-resolution outputs for expressive video scenes.
59
+ [Wan I2V 720P]
60
+ The dataset used consists of 149 videos (each one hand-selected) in 1280x720x96 resolution but was trained in 244p and 480p and 64 frames with 64 dim (L40s).
61
+ Trigger word was used so it needs to be included in the prompt:
62
+ l1v3w4llp4p3r
63
+ [Hunyuan T2V]
64
+ The dataset used consists of 529 videos (each one hand-selected) in 1280x720x96 resolution but was trained in 244p and 72 frames with 64 dim (multiple RTX 4090).
65
+ No captions or activation words were used, the only control you will need to adjust is the lora strength.
66
+ Another important note is that it was trained in full blocks, I don't know how it will behave when mixing 2 or more loras, if you want to mix and are not getting a good result, try disabling single blocks.
67
+ I recommend using lora strength between 0.2 and 1.2 maximum, resolution 1280x720 or generate at 512 and upscale later, minimum 3 seconds (72 frames + 1).
68
+ [LTXV I2V 13b 0.9.7 โ€“ Experimental v1]
69
+ The model was trained on 140 curated videos (512px, 24fps, 49 frames), using 250 epochs, 32 dim, and AdamW8bit.
70
+ It was trained using Diffusion Pipe with support for LTXV I2V v0.9.7 (13B).
71
+ Captions were used and generated with Qwen2.5-VL-7B via a structured prompt format.
72
+ This is an
73
+ experimental first version
74
+ , so expect some variability depending on seed and prompt detail.
75
+ Recommended:
76
+ Scheduler: sgm_uniform
77
+ Sampler: euler
78
+ Steps: 30
79
+ โš ๏ธ Long prompts are highly recommended to avoid motion artifacts.
80
+ You can generate captions using the
81
+ Ollama Describer
82
+ or optionally use the official LTXV Prompt Enhancer.
83
+ For more details, see the
84
+ About this version
85
+ tab.
86
+ ------------------------------------------------------------------------------------------------------
87
+ For more details see the version description
88
+ Share your results.
89
+
90
+ ## Version Notes (LTXV 13b 0.9.7 I2V)
91
+
92
+ ๐Ÿงช Overview
93
+ This is an
94
+ experimental LoRA
95
+ trained for the
96
+ LTXV I2V (Image-to-Video) model
97
+ , version
98
+ 0.9.7
99
+ , with
100
+ 13B parameters
101
+ . It transforms static images into
102
+ fluid, seamless animated loops
103
+ , with natural motion applied only to flexible or dynamic elements โ€” like hair, clothing, particles, and ambient light โ€” while preserving
104
+ rigid structure stability
105
+ (e.g., armor, weapons, mechanical parts).
106
+ This is the
107
+ first version
108
+ of this LoRA, and results may vary depending on the prompt quality and seed. Better versions may be released in the future as training techniques are refined.
109
+ My Workflow
110
+ โš™๏ธ Training Details
111
+ ๐Ÿง 
112
+ Base Model
113
+ : LTXV I2V v0.9.7 (13B parameters)
114
+ ๐ŸŽž๏ธ
115
+ Video Dataset
116
+ : 140 short clips
117
+ โฑ๏ธ
118
+ Frame Rate
119
+ : 24 fps
120
+ ๐Ÿงฎ
121
+ Frames per Video
122
+ : 49
123
+ ๐Ÿ–ผ๏ธ
124
+ Resolution
125
+ : 512px
126
+ ๐Ÿ”
127
+ Epochs
128
+ : 250
129
+ ๐Ÿงฎ
130
+ Total Training Steps
131
+ : ~35,000
132
+ ๐Ÿ“‰
133
+ Learning Rate
134
+ :
135
+ 1e-4
136
+ ๐Ÿ“ฆ
137
+ Batch Size
138
+ : 1
139
+ ๐Ÿ“
140
+ LoRA Dimension
141
+ : 32
142
+ โš™๏ธ
143
+ Optimizer
144
+ :
145
+ AdamW8bit
146
+ ๐Ÿ› ๏ธ
147
+ Trainer Used
148
+ :
149
+ Diffusion Pipe (by tdrussell)
150
+ ๐Ÿšซ
151
+ Not using official trainer
152
+ :
153
+ LTX-Video-Trainer (by Lightricks)
154
+ Layer Coverage
155
+ :
156
+ When trained using
157
+ Diffusion Pipe
158
+ ,
159
+ all layers
160
+ were updated during LoRA training.
161
+ In contrast, the
162
+ official trainer
163
+ from LightTricks (LTX-Video-Trainer) by default
164
+ only updates attention layers
165
+ (e.g.,
166
+ to_k
167
+ ,
168
+ to_q
169
+ ,
170
+ to_v
171
+ ,
172
+ to_out.0
173
+ ), making it possible to use higher
174
+ dim
175
+ (e.g., 128) while still keeping the file size low (~700MB).
176
+ Initial Loss
177
+ : High โ€” LTXV I2V is known to require many steps before reaching stability
178
+ โš ๏ธ The I2V 13B model begins with a
179
+ very high initial loss
180
+ , and convergence is slow โ€” requiring many steps to stabilize below 0.1. Training this architecture is not plug-and-play and takes persistence.
181
+ โš ๏ธ Prompting Recommendations
182
+ This LoRA is
183
+ very sensitive to prompt quality and seed variation
184
+ .
185
+ Using short or unclear prompts often causes:
186
+ Rigid elements
187
+ like weapons or chairs to appear
188
+ soft or rubbery
189
+ Unintended
190
+ motion of static parts
191
+ (e.g., armor bending, background flickering)
192
+ These artifacts are not due to the LoRA itself but rather to a
193
+ lack of motion guidance
194
+ in the prompt or an unsuitable seed.
195
+ โœ… To get the best results:
196
+ Use
197
+ long, detailed prompts
198
+ that clearly separate moving vs non-moving parts
199
+ Try changing the
200
+ seed
201
+ if you're seeing unwanted distortion
202
+ You can generate prompts automatically using my custom ComfyUI node:
203
+ ๐Ÿ”ง
204
+ Ollama Describer
205
+ This node uses a vision-capable LLM to generate motion-aware captions. In my case, I used
206
+ Qwen2.5-VL-7B
207
+ to generate all motion prompts during training and testing.
208
+ ๐Ÿ’ก Alternatively, the
209
+ LTXV Prompt Enhancer
210
+ from Lightricks' custom node set may also be used for prompt conditioning.
211
+ ๐Ÿง  Recommended Prompt Template
212
+ Use this with any vision-enabled LLM like Qwen-VL, Gemini, or GPT-4o:
213
+ You are an expert in motion design for seamless animated loops.
214
+
215
+ Given a single image as input, generate a richly detailed description of how it could be turned into a smooth, seamless animation.
216
+
217
+ Your response must include:
218
+
219
+ โœ… What elements **should move**:
220
+ โ€“ Hair (e.g., swaying, fluttering)
221
+ โ€“ Eyes (e.g., blinking, subtle gaze shifts)
222
+ โ€“ Clothing or fabric elements (e.g., ribbons, loose parts reacting to wind or motion)
223
+ โ€“ Ambient particles (e.g., dust, sparks, petals)
224
+ โ€“ Light effects (e.g., holograms, glows, energy fields)
225
+ โ€“ Floating objects (e.g., drones, magical orbs) if they are clearly not rigid or fixed
226
+ โ€“ Background **ambient** motion (e.g., fog, drifting light, slow parallax)
227
+
228
+ ๐Ÿšซ And **explicitly specify what should remain static**:
229
+ โ€“ Rigid structures (e.g., chairs, weapons, metallic armor)
230
+ โ€“ Body parts not involved in subtle motion (e.g., torso, limbs unless thereโ€™s idle shifting)
231
+ โ€“ Background elements that do not visually suggest movement
232
+
233
+ โš ๏ธ Guidelines:
234
+ โ€“ The animation must be **fluid, consistent, and seamless**, suitable for a loop
235
+ โ€“ Do NOT include sudden movements, teleportation, scene transitions, or pose changes
236
+ โ€“ Do NOT invent objects or effects not present in the image
237
+ โ€“ Do NOT describe static features like colors, names, or environment themes
238
+ โ€“ The output must begin with the trigger word: **lvwpr**
239
+ โ€“ Return only the description (no lists, no markdown, no instructions)
240
+ ๐Ÿงช Experimental Status
241
+ This is the
242
+ first public version
243
+ of this LoRA for LTXV I2V.
244
+ If I discover new training techniques, better captioning strategies, or improvements in convergence,
245
+ future versions will be released
246
+ with higher quality and better performance.
247
+ ๐Ÿ™Œ Feedback Welcome
248
+ If you create something interesting with this LoRA, feel free to share what youโ€™ve made.
249
+ Iโ€™ll be checking community uploads โ€” and if I find your results particularly impressive, Iโ€™ll help give them
250
+ a boost of Civitai buzz
251
+ ๐Ÿ˜‰
252
+
253
+ ---
254
+
255
+ ## Civitai Links
256
+
257
+ * **[๐Ÿ”— View This Version on Civitai โ†’](https://civitai.com/models/1264662?modelVersionId=1853133)**
258
+ * [View Full Model Page โ†’](https://civitai.com/models/1264662)
259
+ * [View Creator Profile โ†’](https://civitai.com/user/NRDX)
260
+
261
+ ---
262
+
263
+ ## Preview Images
264
+
265
+ ![Preview Image](./previews/1853133_79618573.mp4)
266
+
267
+ ---
268
+
269
+ ## File Information
270
+
271
+ * **Filename**: `livewallpaper_e250.safetensors`
272
+ * **Size**: 312.13 MB
273
+ * **Hash (AutoV2)**: `8C1145C11B`
274
+ * **Hash (SHA256)**: `8C1145C11B645300F958D18239767587AFAA0106A85349450FB05DBFA142ACD4`
275
+