Update api/ltx_server_refactored.py
Browse files
api/ltx_server_refactored.py
CHANGED
|
@@ -598,8 +598,8 @@ class VideoService:
|
|
| 598 |
print(f" [LOG] Latentes carregados para a GPU. Shape: {latents_to_refine.shape}")
|
| 599 |
|
| 600 |
with torch.autocast(device_type=self.device.split(':')[0], dtype=self.runtime_autocast_dtype, enabled=(self.device == 'cuda')):
|
| 601 |
-
refine_height = latents_to_refine.shape[3] * self.pipeline.vae_scale_factor
|
| 602 |
-
refine_width = latents_to_refine.shape[4] * self.pipeline.vae_scale_factor
|
| 603 |
second_pass_kwargs = {
|
| 604 |
"prompt": prompt, "negative_prompt": negative_prompt, "height": refine_height, "width": refine_width,
|
| 605 |
"frame_rate": int(DEFAULT_FPS), "num_frames": latents_to_refine.shape[2],
|
|
|
|
| 598 |
print(f" [LOG] Latentes carregados para a GPU. Shape: {latents_to_refine.shape}")
|
| 599 |
|
| 600 |
with torch.autocast(device_type=self.device.split(':')[0], dtype=self.runtime_autocast_dtype, enabled=(self.device == 'cuda')):
|
| 601 |
+
refine_height = latents_to_refine.shape[3] * self.pipeline.vae_scale_factor
|
| 602 |
+
refine_width = latents_to_refine.shape[4] * self.pipeline.vae_scale_factor
|
| 603 |
second_pass_kwargs = {
|
| 604 |
"prompt": prompt, "negative_prompt": negative_prompt, "height": refine_height, "width": refine_width,
|
| 605 |
"frame_rate": int(DEFAULT_FPS), "num_frames": latents_to_refine.shape[2],
|