LTX-2.3 INT8 Quantized Models
I have quantized various LTX-2.3 models and text encoders into INT8 to be used in ComfyUI.
Make sure you have INT8 nodes installed, as ComfyUI does not natively support INT8 as of Q1 2026.
These INT8 models are designed to speed up workflows on Ampere (RTX 30XX) GPUs, which are relatively inefficient with FP8 or FP4 quantization.
Demo Video
Usage
Example loading with Transformers:
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("Winnougan/LTX-2.3-INT8")
model = AutoModelForSeq2SeqLM.from_pretrained("Winnougan/LTX-2.3-INT8")
These models are primarily intended for ComfyUI workflows using INT8-compatible nodes.
Notes
- Quantization performed for performance on Ampere GPUs (RTX 30XX series).
- INT8 offers a good balance between speed, VRAM usage, and quality.
- Ensure your ComfyUI installation supports INT8 tensor loading.
Disclaimer
These are quantized versions of the original LTX-2.3 models.
All credit for the original models goes to their respective authors.
Quantization may introduce minor differences in output quality compared to FP16 models.
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support