Any-to-Any
English
ltx-video
image-to-video
text-to-video
File size: 2,245 Bytes
98328d3
28be962
 
 
 
 
 
98328d3
 
 
28be962
98328d3
 
 
 
 
 
 
 
 
 
 
28be962
 
 
 
 
 
98328d3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28be962
98328d3
 
 
28be962
 
 
 
 
 
 
 
 
 
 
 
98328d3
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
base_model:
- Lightricks/LTX-2
datasets:
- Lightricks/Canny-Control-Dataset
language:
- en
license: other
license_name: ltx-2-community-license
license_link: https://www.github.com/Lightricks/LTX-2/LICENSE
pipeline_tag: any-to-any
tags:
- ltx-video
- image-to-video
- text-to-video
pinned: true
---

# LTX-2 19B IC-LoRA Canny Control

This is a Canny control IC-LoRA trained on top of **LTX-2-19b**, enabling structure-preserving video generation from text and reference frames.

It is based on the [LTX-2](https://huggingface.co/papers/2601.03233) foundation model.

- **Paper:** [LTX-2: Efficient Joint Audio-Visual Foundation Model](https://huggingface.co/papers/2601.03233)
- **Code:** [GitHub Repository](https://github.com/Lightricks/LTX-2)
- **Project Page:** [LTX-2 Playground](https://app.ltx.studio/ltx-2-playground/i2v)

## What is In-Context LoRA (IC LoRA)?

IC LoRA enables conditioning video generation on reference video frames at inference time, allowing fine-grained video-to-video control on top of a text-to-video, base model.
It allows also the usage of an intial image for image-to-video, and generate audio-visual output.

## Model Files

`ltx-2-19b-ic-lora-canny-control.safetensors`

## License

See the **LTX-2-community-license** for full terms.

## Model Details

- **Base Model:** LTX-2-19b Video
- **Training Type:** IC LoRA
- **Control Type:** Canny edge conditioning

### 🔌 Using in ComfyUI
1. Copy the LoRA weights into `models/loras`.
2. Use the official IC-LoRA workflow from the [LTX-2 ComfyUI repository](https://github.com/Lightricks/ComfyUI-LTXVideo/).

## Dataset

The model was trained using the [Lightricks/Canny-Control-Dataset](https://huggingface.co/datasets/Lightricks/Canny-Control-Dataset/).

## Citation

```bibtex
@article{hacohen2025ltx2,
  title={LTX-2: Efficient Joint Audio-Visual Foundation Model},
  author={HaCohen, Yoav and Brazowski, Benny and Chiprut, Nisan and Bitterman, Yaki and Kvochko, Andrew and Berkowitz, Avishai and Shalem, Daniel and Lifschitz, Daphna and Moshe, Dudu and Porat, Eitan and others},
  journal={arXiv preprint arXiv:2601.03233},
  year={2025}
}
```

## Acknowledgments

- Base model by **Lightricks**
- Training infrastructure: **LTX-2 Community Trainer**