scoobyltx / README.md
comfyuiman's picture
Upload folder using huggingface_hub
9ff7ada verified
---
tags:
- ltx-2
- ltx-video
- text-to-video
- audio-video
pinned: true
language:
- en
license: other
pipeline_tag: text-to-video
library_name: diffusers
---
# scoobyltx
This is a fine-tuned version of [`ltx-2-19b-dev.safetensors`](/workspace/models/ltx2/ltx-2-19b-dev.safetensors) trained on custom data.
## Model Details
- **Base Model:** [`ltx-2-19b-dev.safetensors`](/workspace/models/ltx2/ltx-2-19b-dev.safetensors)
- **Training Type:** LoRA fine-tuning
- **Training Steps:** 8000
- **Learning Rate:** 0.0001
- **Batch Size:** 1
## Sample Outputs
| | | | |
|:---:|:---:|:---:|:---:|
## Usage
This model is designed to be used with the LTX-2 (Lightricks Audio-Video) pipeline.
### 🔌 Using Trained LoRAs in ComfyUI
In order to use the trained LoRA in ComfyUI, follow these steps:
1. Copy your trained LoRA checkpoint (`.safetensors` file) to the `models/loras` folder in your ComfyUI installation.
2. In your ComfyUI workflow:
- Add the "Load LoRA" node to choose your LoRA file
- Connect it to the "Load Checkpoint" node to apply the LoRA to the base model
You can find reference Text-to-Video (T2V) and Image-to-Video (I2V) workflows in the
official [LTX-2 repository](https://github.com/Lightricks/LTX-2).
### Example Prompts
This model inherits the license of the base model ([`ltx-2-19b-dev.safetensors`](/workspace/models/ltx2/ltx-2-19b-dev.safetensors)).
## Acknowledgments
- Base model: [Lightricks](https://huggingface.co/Lightricks/LTX-2)
- Trainer: [LTX-2](https://github.com/Lightricks/LTX-2)