Commit
·
4d4ea09
1
Parent(s):
3df02d6
Add README and example videos
Browse filesCo-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
- .gitattributes +1 -0
- AnimateDiff_00774.mp4 +3 -0
- AnimateDiff_00777.mp4 +3 -0
- AnimateDiff_00778.mp4 +3 -0
- README.md +57 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
AnimateDiff_00774.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6644856f28b9b445cd02a9ba9f297e18cc076a7b32c9a9572ab2925fa01a8bd9
|
| 3 |
+
size 5364557
|
AnimateDiff_00777.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e3219f166e7f849e79bb3d3a1c817bc168fdab4cd087598a7c14c75e0542cf13
|
| 3 |
+
size 2537748
|
AnimateDiff_00778.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5ecb6edfe62cd8e34e5f9804e8f69081bf8faec4f757a01e82786c39ece91949
|
| 3 |
+
size 4326717
|
README.md
ADDED
|
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# LTX-2 Image-to-Video Adapter LoRA
|
| 2 |
+
|
| 3 |
+
A high-rank LoRA adapter for [LTX-Video 2](https://github.com/Lightricks/LTX-Video) that substantially improves image-to-video generation quality. No complex workflows, no image preprocessing, no compression tricks -- just a direct image embedding pipeline that works.
|
| 4 |
+
|
| 5 |
+
## What This Is
|
| 6 |
+
|
| 7 |
+
This LoRA was trained on **30,000 generated videos** spanning a wide range of subjects, styles, and motion types. The result is a highly generalized adapter that strengthens LTX-2's ability to take a single image and produce coherent, high-fidelity video from it.
|
| 8 |
+
|
| 9 |
+
### Key Specs
|
| 10 |
+
|
| 11 |
+
| Parameter | Value |
|
| 12 |
+
|-----------|-------|
|
| 13 |
+
| **Base Model** | LTX-Video 2 |
|
| 14 |
+
| **LoRA Rank** | 256 |
|
| 15 |
+
| **Training Set** | ~30,000 generated videos |
|
| 16 |
+
| **Training Scope** | Visual only (no explicit audio training) |
|
| 17 |
+
|
| 18 |
+
## What It Does
|
| 19 |
+
|
| 20 |
+
- **Improved image fidelity** -- the generated video maintains stronger adherence to the source image with less drift or distortion across frames.
|
| 21 |
+
- **Better motion coherence** -- subjects move more naturally and consistently throughout the clip.
|
| 22 |
+
- **Broader generalization** -- performs well across diverse subjects and scenes without needing per-category tuning.
|
| 23 |
+
- **Zero-workflow overhead** -- no ControlNet, no IP-Adapter stacking, no image manipulation required. Load the LoRA, attach an image embedding, prompt, and generate.
|
| 24 |
+
|
| 25 |
+
### A Note on Audio
|
| 26 |
+
|
| 27 |
+
Audio was **not** explicitly trained into this LoRA. However, due to the nature of how LTX-2 handles its latent space, there are subtle shifts in audio output compared to the base model. This is a side effect of the training process, not an intentional feature.
|
| 28 |
+
|
| 29 |
+
## Usage (ComfyUI)
|
| 30 |
+
|
| 31 |
+
1. Place the LoRA file in your `ComfyUI/models/loras/` directory.
|
| 32 |
+
2. Add an **LTX-2** model loader node and load the base LTX-2 checkpoint.
|
| 33 |
+
3. Add a **Load LoRA** node and select this adapter.
|
| 34 |
+
4. Connect an **image embedding** node with your source image.
|
| 35 |
+
5. Add your text prompt and generate.
|
| 36 |
+
|
| 37 |
+
No additional nodes, preprocessing steps, or auxiliary models are needed.
|
| 38 |
+
|
| 39 |
+
## Examples
|
| 40 |
+
|
| 41 |
+
Three reference videos demonstrating the adapter's output quality:
|
| 42 |
+
|
| 43 |
+
<video src="https://huggingface.co/MachineDelusions/LTX-2_Image2Video_Adapter_LoRa/resolve/main/AnimateDiff_00774.mp4" autoplay loop muted playsinline></video>
|
| 44 |
+
|
| 45 |
+
<video src="https://huggingface.co/MachineDelusions/LTX-2_Image2Video_Adapter_LoRa/resolve/main/AnimateDiff_00777.mp4" autoplay loop muted playsinline></video>
|
| 46 |
+
|
| 47 |
+
<video src="https://huggingface.co/MachineDelusions/LTX-2_Image2Video_Adapter_LoRa/resolve/main/AnimateDiff_00778.mp4" autoplay loop muted playsinline></video>
|
| 48 |
+
|
| 49 |
+
## Model Details
|
| 50 |
+
|
| 51 |
+
- **Architecture:** LoRA (Low-Rank Adaptation) applied to LTX-Video 2's transformer layers
|
| 52 |
+
- **Rank 256** provides a high-capacity adaptation while remaining efficient to load and merge
|
| 53 |
+
- **Training data** was intentionally diverse to avoid overfitting to any single domain, producing a general-purpose image-to-video adapter rather than a style-specific fine-tune
|
| 54 |
+
|
| 55 |
+
## License
|
| 56 |
+
|
| 57 |
+
Please refer to the [LTX-Video license](https://github.com/Lightricks/LTX-Video) for base model terms.
|