Softology commited on
Commit
730af89
·
verified ·
1 Parent(s): 2924081

Upload 2 files

Browse files
Lower_VRAM_LTX-2_I2V_Distilled_wLora.json CHANGED
@@ -87,7 +87,7 @@
87
  58
88
  ],
89
  "flags": {},
90
- "order": 13,
91
  "mode": 0,
92
  "inputs": [
93
  {
@@ -130,7 +130,7 @@
130
  78
131
  ],
132
  "flags": {},
133
- "order": 20,
134
  "mode": 0,
135
  "inputs": [
136
  {
@@ -183,7 +183,7 @@
183
  817.489191202842
184
  ],
185
  "flags": {},
186
- "order": 21,
187
  "mode": 0,
188
  "inputs": [
189
  {
@@ -252,7 +252,7 @@
252
  82
253
  ],
254
  "flags": {},
255
- "order": 14,
256
  "mode": 4,
257
  "inputs": [
258
  {
@@ -292,7 +292,7 @@
292
  82
293
  ],
294
  "flags": {},
295
- "order": 15,
296
  "mode": 4,
297
  "inputs": [
298
  {
@@ -334,7 +334,7 @@
334
  "flags": {
335
  "collapsed": true
336
  },
337
- "order": 17,
338
  "mode": 0,
339
  "inputs": [
340
  {
@@ -386,7 +386,7 @@
386
  94
387
  ],
388
  "flags": {},
389
- "order": 18,
390
  "mode": 0,
391
  "inputs": [
392
  {
@@ -468,58 +468,6 @@
468
  "ltx-2-spatial-upscaler-x2-1.0.safetensors"
469
  ]
470
  },
471
- {
472
- "id": 5213,
473
- "type": "MarkdownNote",
474
- "pos": [
475
- -2682.3402563401937,
476
- 6633.478072231158
477
- ],
478
- "size": [
479
- 240.16187366614236,
480
- 98.41687371924309
481
- ],
482
- "flags": {
483
- "collapsed": false
484
- },
485
- "order": 4,
486
- "mode": 0,
487
- "inputs": [],
488
- "outputs": [],
489
- "title": "Video Size",
490
- "properties": {},
491
- "widgets_values": [
492
- "Width & height settings must be divisible by 32 + 1. \nFrame count must be divisible by 8 + 1. \n\n\nRunning with invalid parameters **will not cause errors**. Instead, the flow will silently choose the closest valid parameters. "
493
- ],
494
- "color": "#432",
495
- "bgcolor": "#653"
496
- },
497
- {
498
- "id": 5214,
499
- "type": "MarkdownNote",
500
- "pos": [
501
- -1985.053544501217,
502
- 6321.755714756273
503
- ],
504
- "size": [
505
- 469.5872194653941,
506
- 209.10385276580928
507
- ],
508
- "flags": {
509
- "collapsed": false
510
- },
511
- "order": 5,
512
- "mode": 0,
513
- "inputs": [],
514
- "outputs": [],
515
- "title": "LoRA",
516
- "properties": {},
517
- "widgets_values": [
518
- "# Using LoRAs\n**Ctrl + B to enable**\n\nFor the camera LoRA we recommend setting strength = 1\n\nYou can use LTX [Ready to use camera LoRAs](https://huggingface.co/collections/Lightricks/ltx-2)\n\n\n\n\nUse a second LoRa only when the details are important in the LoRa. \n\n## Training\nWe have made available a full trainer to help you to train your own LoRA. You can [find it here](https://github.com/Lightricks/LTX-2)."
519
- ],
520
- "color": "#432",
521
- "bgcolor": "#653"
522
- },
523
  {
524
  "id": 5192,
525
  "type": "LTXVGemmaEnhancePrompt",
@@ -534,7 +482,7 @@
534
  "flags": {
535
  "collapsed": true
536
  },
537
- "order": 16,
538
  "mode": 0,
539
  "inputs": [
540
  {
@@ -578,7 +526,7 @@
578
  "You are a Creative Assistant writing concise, action-focused image-to-video prompts. Given an image (first frame) and user Raw Input Prompt, generate a prompt to guide video generation from that image.\n\n#### Guidelines:\n- Analyze the Image: Identify Subject, Setting, Elements, Style and Mood.\n- Follow user Raw Input Prompt: Include all requested motion, actions, camera movements, audio, and details. If in conflict with the image, prioritize user request while maintaining visual consistency (describe transition from image to user's scene).\n- Describe only changes from the image: Don't reiterate established visual details. Inaccurate descriptions may cause scene cuts.\n- Active language: Use present-progressive verbs (\"is walking,\" \"speaking\"). If no action specified, describe natural movements.\n- Chronological flow: Use temporal connectors (\"as,\" \"then,\" \"while\").\n- Audio layer: Describe complete soundscape throughout the prompt alongside actions—NOT at the end. Align audio intensity with action tempo. Include natural background audio, ambient sounds, effects, speech or music (when requested). Be specific (e.g., \"soft footsteps on tile\") not vague (e.g., \"ambient sound\").\n- Speech (only when requested): Provide exact words in quotes with character's visual/voice characteristics (e.g., \"The tall man speaks in a low, gravelly voice\"), language if not English and accent if relevant. If general conversation mentioned without text, generate contextual quoted dialogue. (i.e., \"The man is talking\" input -> the output should include exact spoken words, like: \"The man is talking in an excited voice saying: 'You won't believe what I just saw!' His hands gesture expressively as he speaks, eyebrows raised with enthusiasm. The ambient sound of a quiet room underscores his animated speech.\")\n- Style: Include visual style at beginning: \"Style: <style>, <rest of prompt>.\" If unclear, omit to avoid conflicts.\n- Visual and audio only: Describe only what is seen and heard. NO smell, taste, or tactile sensations.\n- Restrained language: Avoid dramatic terms. Use mild, natural, understated phrasing.\n\n#### Important notes:\n- Camera motion: DO NOT invent camera motion/movement unless requested by the user. Make sure to include camera motion only if specified in the input.\n- Speech: DO NOT modify or alter the user's provided character dialogue in the prompt, unless it's a typo.\n- No timestamps or cuts: DO NOT use timestamps or describe scene cuts unless explicitly requested.\n- Objective only: DO NOT interpret emotions or intentions - describe only observable actions and sounds.\n- Format: DO NOT use phrases like \"The scene opens with...\" / \"The video starts...\". Start directly with Style (optional) and chronological scene description.\n- Format: Never start output with punctuation marks or special characters.\n- DO NOT invent dialogue unless the user mentions speech/talking/singing/conversation.\n- Your performance is CRITICAL. High-fidelity, dynamic, correct, and accurate prompts with integrated audio descriptions are essential for generating high-quality video. Your goal is flawless execution of these rules.\n\n#### Output Format (Strict):\n- Single concise paragraph in natural English. NO titles, headings, prefaces, sections, code fences, or Markdown.\n- If unsafe/invalid, return original user prompt. Never ask questions or clarifications.\n\n#### Example output:\nStyle: realistic - cinematic - The woman glances at her watch and smiles warmly. She speaks in a cheerful, friendly voice, \"I think we're right on time!\" In the background, a café barista prepares drinks at the counter. The barista calls out in a clear, upbeat tone, \"Two cappuccinos ready!\" The sound of the espresso machine hissing softly blends with gentle background chatter and the light clinking of cups on saucers.",
579
  512,
580
  false,
581
- 42,
582
  "randomize"
583
  ]
584
  },
@@ -594,7 +542,7 @@
594
  424.68799816112823
595
  ],
596
  "flags": {},
597
- "order": 6,
598
  "mode": 0,
599
  "inputs": [],
600
  "outputs": [
@@ -618,58 +566,6 @@
618
  "color": "#232",
619
  "bgcolor": "#353"
620
  },
621
- {
622
- "id": 5211,
623
- "type": "MarkdownNote",
624
- "pos": [
625
- -3620.0939157608577,
626
- 7287.881055902407
627
- ],
628
- "size": [
629
- 420.9178919734818,
630
- 209.33158232705227
631
- ],
632
- "flags": {
633
- "collapsed": false
634
- },
635
- "order": 7,
636
- "mode": 0,
637
- "inputs": [],
638
- "outputs": [],
639
- "title": "Prompting LTX-2",
640
- "properties": {},
641
- "widgets_values": [
642
- "## LTX-2 Prompting Tips\n1. **Core Actions**: Describe events and actions as they occur over time \n2. **Audio**: Describe sounds and dialogue needed for the scene \n3. **Reference Image**: Do not repeat details already present \n4. **Consistency**: Avoid instructions that do not match the reference image, as this will degrade results"
643
- ],
644
- "color": "#432",
645
- "bgcolor": "#653"
646
- },
647
- {
648
- "id": 5212,
649
- "type": "MarkdownNote",
650
- "pos": [
651
- -3627.789651297241,
652
- 6716.012179425959
653
- ],
654
- "size": [
655
- 427.8617596510994,
656
- 529.3430202216196
657
- ],
658
- "flags": {
659
- "collapsed": false
660
- },
661
- "order": 8,
662
- "mode": 0,
663
- "inputs": [],
664
- "outputs": [],
665
- "title": "Model Links",
666
- "properties": {},
667
- "widgets_values": [
668
- "## Model Links\n\n**LTX-2 Model Weights**\n\n- [ltx-2-19b-distilled.safetensors](https://huggingface.co/Lightricks/LTX-2/resolve/main/ltx-2-19b-distilled.safetensors)\n- [ ltx-2-spatial-upscaler-x2-1.0.safetensors ](https://huggingface.co/Lightricks/LTX-2/resolve/main/ltx-2-spatial-upscaler-x2-1.0.safetensors)\n\n**Text Encoder**\n- [Google Gemma 3](https://huggingface.co/google/gemma-3-12b-it-qat-q4_0-unquantized)\n\nPlease download the entire folder - \n\n- Run: \n 1. cd models/text_encoders\n 2. git clone https://huggingface.co/google/gemma-3-12b-it-qat-q4_0-unquantized\n\n*Full documentation can be found [here](https://docs.ltx.video/open-source-model/integration-tools/comfy-ui#text-encoder)\n\n\n**Model Storage Location**\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 text_encoders/\n│ │ ├── gemma-3-12b-it-qat-q4_0-unquantized\\model-00001-of-00005.safetensors\n│ ├── 📂 checkpoints/\n│ │ └── ltx-2-19b-distilled.safetensors\n│ ├── 📂 latent_upscale_models/ \n └── ltx-2-spatial-upscaler-x2-1.0.safetensors\n```\n**Assets**\n- [distilled image](https://github.com/Lightricks/ComfyUI-LTXVideo/blob/master/example_workflows/assets/distilled%20image.png)\n\n## Report Issues\nTo report any issues when running this workflow, [go to GitHub](https://github.com/Lightricks/ComfyUI-LTXVideo/issues)\n"
669
- ],
670
- "color": "#432",
671
- "bgcolor": "#653"
672
- },
673
  {
674
  "id": 5176,
675
  "type": "CheckpointLoaderSimple",
@@ -682,7 +578,7 @@
682
  115.06840156250018
683
  ],
684
  "flags": {},
685
- "order": 9,
686
  "mode": 0,
687
  "inputs": [],
688
  "outputs": [
@@ -728,7 +624,7 @@
728
  58
729
  ],
730
  "flags": {},
731
- "order": 10,
732
  "mode": 0,
733
  "inputs": [],
734
  "outputs": [
@@ -761,7 +657,7 @@
761
  106
762
  ],
763
  "flags": {},
764
- "order": 11,
765
  "mode": 0,
766
  "inputs": [],
767
  "outputs": [
@@ -802,7 +698,7 @@
802
  433.14022877831576
803
  ],
804
  "flags": {},
805
- "order": 12,
806
  "mode": 0,
807
  "inputs": [],
808
  "outputs": [
@@ -842,7 +738,7 @@
842
  334
843
  ],
844
  "flags": {},
845
- "order": 19,
846
  "mode": 0,
847
  "inputs": [
848
  {
@@ -2790,7 +2686,7 @@
2790
  "Node name for S&R": "RandomNoise"
2791
  },
2792
  "widgets_values": [
2793
- 420,
2794
  "randomize"
2795
  ]
2796
  }
@@ -3269,10 +3165,10 @@
3269
  "VHS_MetadataImage": true,
3270
  "VHS_KeepIntermediate": true,
3271
  "ds": {
3272
- "scale": 0.5684186925550379,
3273
  "offset": [
3274
- 3798.709570601023,
3275
- -6181.509833017723
3276
  ]
3277
  }
3278
  },
 
87
  58
88
  ],
89
  "flags": {},
90
+ "order": 9,
91
  "mode": 0,
92
  "inputs": [
93
  {
 
130
  78
131
  ],
132
  "flags": {},
133
+ "order": 16,
134
  "mode": 0,
135
  "inputs": [
136
  {
 
183
  817.489191202842
184
  ],
185
  "flags": {},
186
+ "order": 17,
187
  "mode": 0,
188
  "inputs": [
189
  {
 
252
  82
253
  ],
254
  "flags": {},
255
+ "order": 10,
256
  "mode": 4,
257
  "inputs": [
258
  {
 
292
  82
293
  ],
294
  "flags": {},
295
+ "order": 11,
296
  "mode": 4,
297
  "inputs": [
298
  {
 
334
  "flags": {
335
  "collapsed": true
336
  },
337
+ "order": 13,
338
  "mode": 0,
339
  "inputs": [
340
  {
 
386
  94
387
  ],
388
  "flags": {},
389
+ "order": 14,
390
  "mode": 0,
391
  "inputs": [
392
  {
 
468
  "ltx-2-spatial-upscaler-x2-1.0.safetensors"
469
  ]
470
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
471
  {
472
  "id": 5192,
473
  "type": "LTXVGemmaEnhancePrompt",
 
482
  "flags": {
483
  "collapsed": true
484
  },
485
+ "order": 12,
486
  "mode": 0,
487
  "inputs": [
488
  {
 
526
  "You are a Creative Assistant writing concise, action-focused image-to-video prompts. Given an image (first frame) and user Raw Input Prompt, generate a prompt to guide video generation from that image.\n\n#### Guidelines:\n- Analyze the Image: Identify Subject, Setting, Elements, Style and Mood.\n- Follow user Raw Input Prompt: Include all requested motion, actions, camera movements, audio, and details. If in conflict with the image, prioritize user request while maintaining visual consistency (describe transition from image to user's scene).\n- Describe only changes from the image: Don't reiterate established visual details. Inaccurate descriptions may cause scene cuts.\n- Active language: Use present-progressive verbs (\"is walking,\" \"speaking\"). If no action specified, describe natural movements.\n- Chronological flow: Use temporal connectors (\"as,\" \"then,\" \"while\").\n- Audio layer: Describe complete soundscape throughout the prompt alongside actions—NOT at the end. Align audio intensity with action tempo. Include natural background audio, ambient sounds, effects, speech or music (when requested). Be specific (e.g., \"soft footsteps on tile\") not vague (e.g., \"ambient sound\").\n- Speech (only when requested): Provide exact words in quotes with character's visual/voice characteristics (e.g., \"The tall man speaks in a low, gravelly voice\"), language if not English and accent if relevant. If general conversation mentioned without text, generate contextual quoted dialogue. (i.e., \"The man is talking\" input -> the output should include exact spoken words, like: \"The man is talking in an excited voice saying: 'You won't believe what I just saw!' His hands gesture expressively as he speaks, eyebrows raised with enthusiasm. The ambient sound of a quiet room underscores his animated speech.\")\n- Style: Include visual style at beginning: \"Style: <style>, <rest of prompt>.\" If unclear, omit to avoid conflicts.\n- Visual and audio only: Describe only what is seen and heard. NO smell, taste, or tactile sensations.\n- Restrained language: Avoid dramatic terms. Use mild, natural, understated phrasing.\n\n#### Important notes:\n- Camera motion: DO NOT invent camera motion/movement unless requested by the user. Make sure to include camera motion only if specified in the input.\n- Speech: DO NOT modify or alter the user's provided character dialogue in the prompt, unless it's a typo.\n- No timestamps or cuts: DO NOT use timestamps or describe scene cuts unless explicitly requested.\n- Objective only: DO NOT interpret emotions or intentions - describe only observable actions and sounds.\n- Format: DO NOT use phrases like \"The scene opens with...\" / \"The video starts...\". Start directly with Style (optional) and chronological scene description.\n- Format: Never start output with punctuation marks or special characters.\n- DO NOT invent dialogue unless the user mentions speech/talking/singing/conversation.\n- Your performance is CRITICAL. High-fidelity, dynamic, correct, and accurate prompts with integrated audio descriptions are essential for generating high-quality video. Your goal is flawless execution of these rules.\n\n#### Output Format (Strict):\n- Single concise paragraph in natural English. NO titles, headings, prefaces, sections, code fences, or Markdown.\n- If unsafe/invalid, return original user prompt. Never ask questions or clarifications.\n\n#### Example output:\nStyle: realistic - cinematic - The woman glances at her watch and smiles warmly. She speaks in a cheerful, friendly voice, \"I think we're right on time!\" In the background, a café barista prepares drinks at the counter. The barista calls out in a clear, upbeat tone, \"Two cappuccinos ready!\" The sound of the espresso machine hissing softly blends with gentle background chatter and the light clinking of cups on saucers.",
527
  512,
528
  false,
529
+ 2367809427,
530
  "randomize"
531
  ]
532
  },
 
542
  424.68799816112823
543
  ],
544
  "flags": {},
545
+ "order": 4,
546
  "mode": 0,
547
  "inputs": [],
548
  "outputs": [
 
566
  "color": "#232",
567
  "bgcolor": "#353"
568
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
569
  {
570
  "id": 5176,
571
  "type": "CheckpointLoaderSimple",
 
578
  115.06840156250018
579
  ],
580
  "flags": {},
581
+ "order": 5,
582
  "mode": 0,
583
  "inputs": [],
584
  "outputs": [
 
624
  58
625
  ],
626
  "flags": {},
627
+ "order": 6,
628
  "mode": 0,
629
  "inputs": [],
630
  "outputs": [
 
657
  106
658
  ],
659
  "flags": {},
660
+ "order": 7,
661
  "mode": 0,
662
  "inputs": [],
663
  "outputs": [
 
698
  433.14022877831576
699
  ],
700
  "flags": {},
701
+ "order": 8,
702
  "mode": 0,
703
  "inputs": [],
704
  "outputs": [
 
738
  334
739
  ],
740
  "flags": {},
741
+ "order": 15,
742
  "mode": 0,
743
  "inputs": [
744
  {
 
2686
  "Node name for S&R": "RandomNoise"
2687
  },
2688
  "widgets_values": [
2689
+ 853453849010343,
2690
  "randomize"
2691
  ]
2692
  }
 
3165
  "VHS_MetadataImage": true,
3166
  "VHS_KeepIntermediate": true,
3167
  "ds": {
3168
+ "scale": 0.687786617991596,
3169
  "offset": [
3170
+ 3301.015187391492,
3171
+ -6404.839993521725
3172
  ]
3173
  }
3174
  },
Lower_VRAM_LTX-2_T2V_Distilled_wLora.json CHANGED
@@ -87,7 +87,7 @@
87
  58
88
  ],
89
  "flags": {},
90
- "order": 12,
91
  "mode": 0,
92
  "inputs": [
93
  {
@@ -166,7 +166,7 @@
166
  78
167
  ],
168
  "flags": {},
169
- "order": 19,
170
  "mode": 0,
171
  "inputs": [
172
  {
@@ -221,7 +221,7 @@
221
  "flags": {
222
  "collapsed": false
223
  },
224
- "order": 17,
225
  "mode": 0,
226
  "inputs": [
227
  {
@@ -282,7 +282,7 @@
282
  90.32621827319144
283
  ],
284
  "flags": {},
285
- "order": 13,
286
  "mode": 4,
287
  "inputs": [
288
  {
@@ -322,7 +322,7 @@
322
  82
323
  ],
324
  "flags": {},
325
- "order": 14,
326
  "mode": 4,
327
  "inputs": [
328
  {
@@ -364,7 +364,7 @@
364
  "flags": {
365
  "collapsed": true
366
  },
367
- "order": 16,
368
  "mode": 0,
369
  "inputs": [
370
  {
@@ -416,7 +416,7 @@
416
  584.6954036515563
417
  ],
418
  "flags": {},
419
- "order": 20,
420
  "mode": 0,
421
  "inputs": [
422
  {
@@ -470,32 +470,6 @@
470
  "ltx-2-spatial-upscaler-x2-1.0.safetensors"
471
  ]
472
  },
473
- {
474
- "id": 5247,
475
- "type": "MarkdownNote",
476
- "pos": [
477
- -2848.6014542681573,
478
- 7401.672050450482
479
- ],
480
- "size": [
481
- 240.16187366614236,
482
- 98.41687371924309
483
- ],
484
- "flags": {
485
- "collapsed": false
486
- },
487
- "order": 4,
488
- "mode": 0,
489
- "inputs": [],
490
- "outputs": [],
491
- "title": "Video Size",
492
- "properties": {},
493
- "widgets_values": [
494
- "Width & height settings must be divisible by 32 + 1. \nFrame count must be divisible by 8 + 1. \n\n\nRunning with invalid parameters **will not cause errors**. Instead, the flow will silently choose the closest valid parameters. "
495
- ],
496
- "color": "#432",
497
- "bgcolor": "#653"
498
- },
499
  {
500
  "id": 5225,
501
  "type": "LTXVGemmaEnhancePrompt",
@@ -510,7 +484,7 @@
510
  "flags": {
511
  "collapsed": true
512
  },
513
- "order": 15,
514
  "mode": 0,
515
  "inputs": [
516
  {
@@ -544,9 +518,10 @@
544
  ],
545
  "title": "Enhancer",
546
  "properties": {
547
- "cnr_id": "ComfyUI-LTXVideo",
548
  "ver": "6790484426305443104ea47d7a2aab13badd8c49",
549
- "Node name for S&R": "LTXVGemmaEnhancePrompt"
 
550
  },
551
  "widgets_values": [
552
  "",
@@ -557,84 +532,6 @@
557
  "randomize"
558
  ]
559
  },
560
- {
561
- "id": 5248,
562
- "type": "MarkdownNote",
563
- "pos": [
564
- -2581.888874206601,
565
- 7104.237854538457
566
- ],
567
- "size": [
568
- 469.5872194653941,
569
- 209.10385276580928
570
- ],
571
- "flags": {
572
- "collapsed": false
573
- },
574
- "order": 5,
575
- "mode": 0,
576
- "inputs": [],
577
- "outputs": [],
578
- "title": "LoRA",
579
- "properties": {},
580
- "widgets_values": [
581
- "# Using LoRAs\n**Ctrl + B to enable**\n\nFor the camera LoRA we recommend setting strength = 1\n\nYou can use LTX [Ready to use camera LoRAs](https://huggingface.co/collections/Lightricks/ltx-2)\n\n\n\n\nUse a second LoRa only when the details are important in the LoRa. \n\n## Training\nWe have made available a full trainer to help you to train your own LoRA. You can [find it here](https://github.com/Lightricks/LTX-2)"
582
- ],
583
- "color": "#432",
584
- "bgcolor": "#653"
585
- },
586
- {
587
- "id": 5246,
588
- "type": "MarkdownNote",
589
- "pos": [
590
- -3759.009376983867,
591
- 7951.430451540103
592
- ],
593
- "size": [
594
- 418.0950915480007,
595
- 208.54998853542747
596
- ],
597
- "flags": {
598
- "collapsed": false
599
- },
600
- "order": 6,
601
- "mode": 0,
602
- "inputs": [],
603
- "outputs": [],
604
- "title": "Prompting LTX-2",
605
- "properties": {},
606
- "widgets_values": [
607
- "## LTX-2 Prompting Tips\n1. **Core Actions**: Describe events and actions as they occur over time \n2. **Audio**: Describe sounds and dialogue needed for the scene \n3. **Consistency**: Avoid instructions that do not match the reference image, as this will degrade results\n"
608
- ],
609
- "color": "#432",
610
- "bgcolor": "#653"
611
- },
612
- {
613
- "id": 5245,
614
- "type": "MarkdownNote",
615
- "pos": [
616
- -3759.592503870824,
617
- 7426.935420507221
618
- ],
619
- "size": [
620
- 419.9816103344465,
621
- 480.7830408486252
622
- ],
623
- "flags": {
624
- "collapsed": false
625
- },
626
- "order": 7,
627
- "mode": 0,
628
- "inputs": [],
629
- "outputs": [],
630
- "title": "Model Links",
631
- "properties": {},
632
- "widgets_values": [
633
- "## Model Links\n\n**LTX-2 Model Weights**\n\n- [ltx-2-19b-distilled.safetensors](https://huggingface.co/Lightricks/LTX-2/resolve/main/ltx-2-19b-distilled.safetensors)\n- [ ltx-2-spatial-upscaler-x2-1.0.safetensors ](https://huggingface.co/Lightricks/LTX-2/resolve/main/ltx-2-spatial-upscaler-x2-1.0.safetensors)\n\n**Text Encoder**\n- [Google Gemma 3](https://huggingface.co/google/gemma-3-12b-it-qat-q4_0-unquantized)\n\nPlease download the entire folder - \n\n- Run: \n 1. cd models/text_encoders\n 2. git clone https://huggingface.co/google/gemma-3-12b-it-qat-q4_0-unquantized\n\n*Full documentation can be found [here](https://docs.ltx.video/open-source-model/integration-tools/comfy-ui#text-encoder)\n\n**Model Storage Location**\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 text_encoders/\n│ │ ├── gemma-3-12b-it-qat-q4_0-unquantized\\model-00001-of-00005.safetensors\n│ ├── 📂 checkpoints/\n│ │ └── ltx-2-19b-distilled.safetensors\n│ ├── 📂 latent_upscale_models/ \n └── ltx-2-spatial-upscaler-x2-1.0.safetensors\n```\n\n## Report Issues\nTo report any issues when running this workflow, [go to GitHub](https://github.com/Lightricks/ComfyUI-LTXVideo/issues)\n"
634
- ],
635
- "color": "#432",
636
- "bgcolor": "#653"
637
- },
638
  {
639
  "id": 5228,
640
  "type": "CheckpointLoaderSimple",
@@ -647,7 +544,7 @@
647
  99.93503409866116
648
  ],
649
  "flags": {},
650
- "order": 8,
651
  "mode": 0,
652
  "inputs": [],
653
  "outputs": [
@@ -693,7 +590,7 @@
693
  106
694
  ],
695
  "flags": {},
696
- "order": 9,
697
  "mode": 0,
698
  "inputs": [],
699
  "outputs": [
@@ -707,9 +604,10 @@
707
  }
708
  ],
709
  "properties": {
710
- "cnr_id": "ComfyUI-LTXVideo",
711
  "ver": "6790484426305443104ea47d7a2aab13badd8c49",
712
  "Node name for S&R": "LTXVGemmaCLIPModelLoader",
 
713
  "widget_ue_connectable": {
714
  "model_path": true,
715
  "max_length": true
@@ -733,7 +631,7 @@
733
  58
734
  ],
735
  "flags": {},
736
- "order": 10,
737
  "mode": 0,
738
  "inputs": [],
739
  "outputs": [
@@ -766,7 +664,7 @@
766
  515.3418506718763
767
  ],
768
  "flags": {},
769
- "order": 11,
770
  "mode": 0,
771
  "inputs": [],
772
  "outputs": [
@@ -802,7 +700,7 @@
802
  290
803
  ],
804
  "flags": {},
805
- "order": 18,
806
  "mode": 0,
807
  "inputs": [
808
  {
@@ -2280,9 +2178,10 @@
2280
  }
2281
  ],
2282
  "properties": {
2283
- "cnr_id": "ComfyUI-LTXVideo",
2284
  "ver": "6790484426305443104ea47d7a2aab13badd8c49",
2285
- "Node name for S&R": "LTXVSpatioTemporalTiledVAEDecode"
 
2286
  },
2287
  "widgets_values": [
2288
  4,
@@ -2815,10 +2714,10 @@
2815
  "extra": {
2816
  "workflowRendererVersion": "LG",
2817
  "ds": {
2818
- "scale": 0.6456703819915997,
2819
  "offset": [
2820
- 3938.7808851325717,
2821
- -6921.75014385692
2822
  ]
2823
  },
2824
  "frontendVersion": "1.36.14",
 
87
  58
88
  ],
89
  "flags": {},
90
+ "order": 8,
91
  "mode": 0,
92
  "inputs": [
93
  {
 
166
  78
167
  ],
168
  "flags": {},
169
+ "order": 15,
170
  "mode": 0,
171
  "inputs": [
172
  {
 
221
  "flags": {
222
  "collapsed": false
223
  },
224
+ "order": 13,
225
  "mode": 0,
226
  "inputs": [
227
  {
 
282
  90.32621827319144
283
  ],
284
  "flags": {},
285
+ "order": 9,
286
  "mode": 4,
287
  "inputs": [
288
  {
 
322
  82
323
  ],
324
  "flags": {},
325
+ "order": 10,
326
  "mode": 4,
327
  "inputs": [
328
  {
 
364
  "flags": {
365
  "collapsed": true
366
  },
367
+ "order": 12,
368
  "mode": 0,
369
  "inputs": [
370
  {
 
416
  584.6954036515563
417
  ],
418
  "flags": {},
419
+ "order": 16,
420
  "mode": 0,
421
  "inputs": [
422
  {
 
470
  "ltx-2-spatial-upscaler-x2-1.0.safetensors"
471
  ]
472
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
473
  {
474
  "id": 5225,
475
  "type": "LTXVGemmaEnhancePrompt",
 
484
  "flags": {
485
  "collapsed": true
486
  },
487
+ "order": 11,
488
  "mode": 0,
489
  "inputs": [
490
  {
 
518
  ],
519
  "title": "Enhancer",
520
  "properties": {
521
+ "aux_id": "Lightricks/ComfyUI-LTXVideo",
522
  "ver": "6790484426305443104ea47d7a2aab13badd8c49",
523
+ "Node name for S&R": "LTXVGemmaEnhancePrompt",
524
+ "cnr_id": "ComfyUI-LTXVideo"
525
  },
526
  "widgets_values": [
527
  "",
 
532
  "randomize"
533
  ]
534
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
535
  {
536
  "id": 5228,
537
  "type": "CheckpointLoaderSimple",
 
544
  99.93503409866116
545
  ],
546
  "flags": {},
547
+ "order": 4,
548
  "mode": 0,
549
  "inputs": [],
550
  "outputs": [
 
590
  106
591
  ],
592
  "flags": {},
593
+ "order": 5,
594
  "mode": 0,
595
  "inputs": [],
596
  "outputs": [
 
604
  }
605
  ],
606
  "properties": {
607
+ "aux_id": "Lightricks/ComfyUI-LTXVideo",
608
  "ver": "6790484426305443104ea47d7a2aab13badd8c49",
609
  "Node name for S&R": "LTXVGemmaCLIPModelLoader",
610
+ "cnr_id": "ComfyUI-LTXVideo",
611
  "widget_ue_connectable": {
612
  "model_path": true,
613
  "max_length": true
 
631
  58
632
  ],
633
  "flags": {},
634
+ "order": 6,
635
  "mode": 0,
636
  "inputs": [],
637
  "outputs": [
 
664
  515.3418506718763
665
  ],
666
  "flags": {},
667
+ "order": 7,
668
  "mode": 0,
669
  "inputs": [],
670
  "outputs": [
 
700
  290
701
  ],
702
  "flags": {},
703
+ "order": 14,
704
  "mode": 0,
705
  "inputs": [
706
  {
 
2178
  }
2179
  ],
2180
  "properties": {
2181
+ "aux_id": "Lightricks/ComfyUI-LTXVideo",
2182
  "ver": "6790484426305443104ea47d7a2aab13badd8c49",
2183
+ "Node name for S&R": "LTXVSpatioTemporalTiledVAEDecode",
2184
+ "cnr_id": "ComfyUI-LTXVideo"
2185
  },
2186
  "widgets_values": [
2187
  4,
 
2714
  "extra": {
2715
  "workflowRendererVersion": "LG",
2716
  "ds": {
2717
+ "scale": 0.7812611622098357,
2718
  "offset": [
2719
+ 3456.4518139711968,
2720
+ -7133.523131013761
2721
  ]
2722
  },
2723
  "frontendVersion": "1.36.14",