Audio
Browse files
app.py
CHANGED
|
@@ -165,6 +165,7 @@ from dwpose import DwposeDetector
|
|
| 165 |
|
| 166 |
|
| 167 |
input_image_debug_value = [None]
|
|
|
|
| 168 |
input_video_debug_value = [None]
|
| 169 |
prompt_debug_value = [None]
|
| 170 |
total_second_length_debug_value = [None]
|
|
@@ -1099,8 +1100,9 @@ class CameraDropdown(gr.HTML):
|
|
| 1099 |
|
| 1100 |
def generate_video_example(input_image, prompt, camera_lora, resolution, radioanimated_mode, input_video, input_audio, progress=gr.Progress(track_tqdm=True)):
|
| 1101 |
|
| 1102 |
-
if input_image_debug_value[0] is not None or
|
| 1103 |
input_image = input_image_debug_value[0]
|
|
|
|
| 1104 |
input_video = input_video_debug_value[0]
|
| 1105 |
prompt = prompt_debug_value[0]
|
| 1106 |
duration = total_second_length_debug_value[0]
|
|
@@ -2340,6 +2342,7 @@ with gr.Blocks(title="LTX-2 Video Distilled 🎥🔈") as demo:
|
|
| 2340 |
)
|
| 2341 |
prompt_debug=gr.Textbox(label="Prompt Debug")
|
| 2342 |
input_image_debug=gr.Image(type="filepath", label="Image Debug")
|
|
|
|
| 2343 |
input_video_debug=gr.Video(label="Video Debug")
|
| 2344 |
total_second_length_debug=gr.Slider(label="Duration Debug", minimum=1, maximum=120, value=5, step=0.1)
|
| 2345 |
resolution_debug = gr.Dropdown(
|
|
@@ -2356,6 +2359,7 @@ with gr.Blocks(title="LTX-2 Video Distilled 🎥🔈") as demo:
|
|
| 2356 |
|
| 2357 |
def handle_field_debug_change(
|
| 2358 |
input_image_debug_data,
|
|
|
|
| 2359 |
input_video_debug_data,
|
| 2360 |
prompt_debug_data,
|
| 2361 |
total_second_length_debug_data,
|
|
@@ -2364,6 +2368,7 @@ with gr.Blocks(title="LTX-2 Video Distilled 🎥🔈") as demo:
|
|
| 2364 |
allocation_time_debug_data
|
| 2365 |
):
|
| 2366 |
input_image_debug_value[0] = input_image_debug_data
|
|
|
|
| 2367 |
input_video_debug_value[0] = input_video_debug_data
|
| 2368 |
prompt_debug_value[0] = prompt_debug_data
|
| 2369 |
total_second_length_debug_value[0] = total_second_length_debug_data
|
|
@@ -2372,9 +2377,10 @@ with gr.Blocks(title="LTX-2 Video Distilled 🎥🔈") as demo:
|
|
| 2372 |
allocation_time_debug_value[0] = allocation_time_debug_data
|
| 2373 |
return []
|
| 2374 |
|
| 2375 |
-
inputs_debug=[input_image_debug, input_video_debug, prompt_debug, total_second_length_debug, resolution_debug, factor_debug, allocation_time_debug]
|
| 2376 |
|
| 2377 |
input_image_debug.upload(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|
|
|
|
| 2378 |
input_video_debug.upload(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|
| 2379 |
prompt_debug.change(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|
| 2380 |
total_second_length_debug.change(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|
|
|
|
| 165 |
|
| 166 |
|
| 167 |
input_image_debug_value = [None]
|
| 168 |
+
input_audio_debug_value = [None]
|
| 169 |
input_video_debug_value = [None]
|
| 170 |
prompt_debug_value = [None]
|
| 171 |
total_second_length_debug_value = [None]
|
|
|
|
| 1100 |
|
| 1101 |
def generate_video_example(input_image, prompt, camera_lora, resolution, radioanimated_mode, input_video, input_audio, progress=gr.Progress(track_tqdm=True)):
|
| 1102 |
|
| 1103 |
+
if input_image_debug_value[0] is not None or input_audio_debug_value[0] is not None or input_video_debug_value[0] is not None or prompt_debug_value[0] is not None or total_second_length_debug_value[0] is not None or allocation_time_debug_value[0] is not None or resolution_debug_value[0] is not None or factor_debug_value[0] is not None:
|
| 1104 |
input_image = input_image_debug_value[0]
|
| 1105 |
+
input_audio = input_audio_debug_value[0]
|
| 1106 |
input_video = input_video_debug_value[0]
|
| 1107 |
prompt = prompt_debug_value[0]
|
| 1108 |
duration = total_second_length_debug_value[0]
|
|
|
|
| 2342 |
)
|
| 2343 |
prompt_debug=gr.Textbox(label="Prompt Debug")
|
| 2344 |
input_image_debug=gr.Image(type="filepath", label="Image Debug")
|
| 2345 |
+
input_audio_debug = gr.Audio(label="Audio Debug", type="filepath")
|
| 2346 |
input_video_debug=gr.Video(label="Video Debug")
|
| 2347 |
total_second_length_debug=gr.Slider(label="Duration Debug", minimum=1, maximum=120, value=5, step=0.1)
|
| 2348 |
resolution_debug = gr.Dropdown(
|
|
|
|
| 2359 |
|
| 2360 |
def handle_field_debug_change(
|
| 2361 |
input_image_debug_data,
|
| 2362 |
+
input_audio_debug_data,
|
| 2363 |
input_video_debug_data,
|
| 2364 |
prompt_debug_data,
|
| 2365 |
total_second_length_debug_data,
|
|
|
|
| 2368 |
allocation_time_debug_data
|
| 2369 |
):
|
| 2370 |
input_image_debug_value[0] = input_image_debug_data
|
| 2371 |
+
input_audio_debug_value[0] = input_audio_debug_data
|
| 2372 |
input_video_debug_value[0] = input_video_debug_data
|
| 2373 |
prompt_debug_value[0] = prompt_debug_data
|
| 2374 |
total_second_length_debug_value[0] = total_second_length_debug_data
|
|
|
|
| 2377 |
allocation_time_debug_value[0] = allocation_time_debug_data
|
| 2378 |
return []
|
| 2379 |
|
| 2380 |
+
inputs_debug=[input_image_debug, input_audio_debug, input_video_debug, prompt_debug, total_second_length_debug, resolution_debug, factor_debug, allocation_time_debug]
|
| 2381 |
|
| 2382 |
input_image_debug.upload(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|
| 2383 |
+
input_audio_debug.upload(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|
| 2384 |
input_video_debug.upload(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|
| 2385 |
prompt_debug.change(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|
| 2386 |
total_second_length_debug.change(fn=handle_field_debug_change, inputs=inputs_debug, outputs=[])
|