ofirbibi commited on
Commit
ebcd621
·
verified ·
1 Parent(s): 5dc10ac

Docs: update readme.

Browse files
Files changed (1) hide show
  1. README.md +3 -7
README.md CHANGED
@@ -38,7 +38,7 @@ demo: https://app.ltx.studio/ltx-2-playground/i2v
38
 
39
  # LTX-2.3 FP8 Model Card
40
 
41
- This is the FP8 versions of the LTX-2.3 model. All information below is derived from the base model.
42
 
43
  This model card focuses on the LTX-2.3 model, which is a significant update to the [LTX-2 model](https://huggingface.co/Lightricks/LTX-2) with improved audio and visual quality as well as enhanced prompt adherence.
44
  LTX-2 was presented in the paper [LTX-2: Efficient Joint Audio-Visual Foundation Model](https://huggingface.co/papers/2601.03233).
@@ -47,7 +47,7 @@ LTX-2 was presented in the paper [LTX-2: Efficient Joint Audio-Visual Foundation
47
 
48
  LTX-2.3 is a DiT-based audio-video foundation model designed to generate synchronized video and audio within a single model. It brings together the core building blocks of modern video generation, with open weights and a focus on practical, local execution.
49
 
50
- [![LTX-2 Open Source](ltx2.3-open.png)](https://youtu.be/o-7us-BR_gQ)
51
 
52
  # Model Checkpoints
53
 
@@ -112,11 +112,7 @@ LTX-2.3 support in the [Diffusers Python library](https://huggingface.co/docs/di
112
 
113
  # Train the model
114
 
115
- The base (dev) model is fully trainable.
116
-
117
- It's extremely easy to reproduce the LoRAs and IC-LoRAs we publish with the model by following the instructions on the [LTX-2 Trainer Readme](https://github.com/Lightricks/LTX-2/blob/main/packages/ltx-trainer/README.md).
118
-
119
- Training for motion, style or likeness (sound+appearance) can take less than an hour in many settings.
120
 
121
  ## Citation
122
 
 
38
 
39
  # LTX-2.3 FP8 Model Card
40
 
41
+ **This is the FP8 versions of the LTX-2.3 model. All information below is derived from the base model.**
42
 
43
  This model card focuses on the LTX-2.3 model, which is a significant update to the [LTX-2 model](https://huggingface.co/Lightricks/LTX-2) with improved audio and visual quality as well as enhanced prompt adherence.
44
  LTX-2 was presented in the paper [LTX-2: Efficient Joint Audio-Visual Foundation Model](https://huggingface.co/papers/2601.03233).
 
47
 
48
  LTX-2.3 is a DiT-based audio-video foundation model designed to generate synchronized video and audio within a single model. It brings together the core building blocks of modern video generation, with open weights and a focus on practical, local execution.
49
 
50
+ [![LTX-2.3 Open Source](https://img.youtube.com/vi/o-7us-BR_gQ/maxresdefault.jpg)](https://youtu.be/o-7us-BR_gQ)
51
 
52
  # Model Checkpoints
53
 
 
112
 
113
  # Train the model
114
 
115
+ Currently it is recommended to train the bf16 model. Recipes for training the fp8 model are welcome as community contributions.
 
 
 
 
116
 
117
  ## Citation
118