Spaces:
Sleeping
Sleeping
check prompt non empty
Browse files
app.py
CHANGED
|
@@ -58,7 +58,9 @@ def start_over(gallery_state, loaded_model_setup):
|
|
| 58 |
return gallery_state, None, None, gr.update(visible=False), loaded_model_setup
|
| 59 |
|
| 60 |
def setup_model(prompt, model, seed, num_iterations, learning_rate, hps_w, imgrw_w, pcks_w, clip_w, progress=gr.Progress(track_tqdm=True)):
|
| 61 |
-
|
|
|
|
|
|
|
| 62 |
"""Clear CUDA memory before starting the training."""
|
| 63 |
torch.cuda.empty_cache() # Free up cached memory
|
| 64 |
|
|
@@ -151,7 +153,7 @@ def generate_image(setup_args, num_iterations):
|
|
| 151 |
|
| 152 |
if error_status["error_occurred"]:
|
| 153 |
torch.cuda.empty_cache() # Free up cached memory
|
| 154 |
-
yield (None, "CUDA out of memory.
|
| 155 |
else:
|
| 156 |
main_thread.join() # Ensure thread completion
|
| 157 |
final_image_path = os.path.join(save_dir, "best_image.png")
|
|
|
|
| 58 |
return gallery_state, None, None, gr.update(visible=False), loaded_model_setup
|
| 59 |
|
| 60 |
def setup_model(prompt, model, seed, num_iterations, learning_rate, hps_w, imgrw_w, pcks_w, clip_w, progress=gr.Progress(track_tqdm=True)):
|
| 61 |
+
if prompt is None:
|
| 62 |
+
raise gr.Error("You forgot to provide a prompt !")
|
| 63 |
+
|
| 64 |
"""Clear CUDA memory before starting the training."""
|
| 65 |
torch.cuda.empty_cache() # Free up cached memory
|
| 66 |
|
|
|
|
| 153 |
|
| 154 |
if error_status["error_occurred"]:
|
| 155 |
torch.cuda.empty_cache() # Free up cached memory
|
| 156 |
+
yield (None, "CUDA out of memory.", None)
|
| 157 |
else:
|
| 158 |
main_thread.join() # Ensure thread completion
|
| 159 |
final_image_path = os.path.join(save_dir, "best_image.png")
|