Update app.py
Browse files
app.py
CHANGED
|
@@ -145,16 +145,20 @@ class App:
|
|
| 145 |
|
| 146 |
with gr.Accordion("Background Music Remover Filter", open=False):
|
| 147 |
cb_bgm_separation = gr.Checkbox(label="Enable Background Music Remover Filter", value=uvr_params["is_separate_bgm"],
|
| 148 |
-
|
| 149 |
-
|
| 150 |
-
dd_uvr_device = gr.Dropdown(label="Device",
|
| 151 |
-
|
|
|
|
|
|
|
| 152 |
dd_uvr_model_size = gr.Dropdown(label="Model", value=uvr_params["model_size"],
|
| 153 |
choices=self.whisper_inf.music_separator.available_models)
|
| 154 |
-
nb_uvr_segment_size = gr.Number(label="Segment Size", value=uvr_params["segment_size"], precision=0
|
| 155 |
-
|
| 156 |
-
|
| 157 |
-
|
|
|
|
|
|
|
| 158 |
|
| 159 |
with gr.Accordion("Advanced processing options", open=False, visible=False):
|
| 160 |
nb_beam_size = gr.Number(label="Beam Size", value=whisper_params["beam_size"], precision=0, interactive=True,
|
|
|
|
| 145 |
|
| 146 |
with gr.Accordion("Background Music Remover Filter", open=False):
|
| 147 |
cb_bgm_separation = gr.Checkbox(label="Enable Background Music Remover Filter", value=uvr_params["is_separate_bgm"],
|
| 148 |
+
info="Enabling this will remove background music by submodel before transcribing",
|
| 149 |
+
interactive=True)
|
| 150 |
+
dd_uvr_device = gr.Dropdown(label="Device",
|
| 151 |
+
value=self.whisper_inf.music_separator.device,
|
| 152 |
+
choices=self.whisper_inf.music_separator.available_devices,
|
| 153 |
+
interactive=True, visible=False)
|
| 154 |
dd_uvr_model_size = gr.Dropdown(label="Model", value=uvr_params["model_size"],
|
| 155 |
choices=self.whisper_inf.music_separator.available_models)
|
| 156 |
+
nb_uvr_segment_size = gr.Number(label="Segment Size", value=uvr_params["segment_size"], precision=0,
|
| 157 |
+
interactive=True, visible=False)
|
| 158 |
+
cb_uvr_save_file = gr.Checkbox(label="Save separated files to output", value=uvr_params["save_file"],
|
| 159 |
+
interactive=True, visible=False)
|
| 160 |
+
cb_uvr_enable_offload = gr.Checkbox(label="Offload sub model after removing background music",value=uvr_params["enable_offload"],
|
| 161 |
+
interactive=True, visible=False)
|
| 162 |
|
| 163 |
with gr.Accordion("Advanced processing options", open=False, visible=False):
|
| 164 |
nb_beam_size = gr.Number(label="Beam Size", value=whisper_params["beam_size"], precision=0, interactive=True,
|