SaltProphet commited on
Commit
20d1efa
·
verified ·
1 Parent(s): 01d791e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +108 -173
app.py CHANGED
@@ -1,4 +1,5 @@
1
  # 1. Install all necessary libraries for the full application
 
2
 
3
  # 2. Import libraries
4
  import gradio as gr
@@ -20,210 +21,144 @@ matplotlib.use('Agg') # Use a non-interactive backend for plotting
20
 
21
  def update_output_visibility(choice):
22
  if "2 Stems" in choice:
23
- return {
24
- vocals_output: gr.update(visible=True),
25
- drums_output: gr.update(visible=False),
26
- bass_output: gr.update(visible=False),
27
- other_output: gr.update(visible=True, label="Instrumental (No Vocals)")
28
- }
29
  elif "4 Stems" in choice:
30
- return {
31
- vocals_output: gr.update(visible=True),
32
- drums_output: gr.update(visible=True),
33
- bass_output: gr.update(visible=True),
34
- other_output: gr.update(visible=True, label="Other")
35
- }
36
 
37
  async def separate_stems(audio_file_path, stem_choice, progress=gr.Progress(track_tqdm=True)):
38
  if audio_file_path is None: raise gr.Error("No audio file uploaded!")
39
- progress(0, desc="Starting...")
40
  try:
41
- progress(0.05, desc="Preparing audio file...")
42
- original_filename_base = os.path.basename(audio_file_path).rsplit('.', 1)[0]
43
- stable_input_path = f"stable_input_{original_filename_base}.wav"
44
- shutil.copy(audio_file_path, stable_input_path)
45
-
46
- model_arg = "--two-stems=vocals" if "2 Stems" in stem_choice else ""
47
- output_dir = "separated"
48
  if os.path.exists(output_dir): shutil.rmtree(output_dir)
49
-
50
  command = f"python3 -m demucs {model_arg} -o \"{output_dir}\" \"{stable_input_path}\""
51
- progress(0.2, desc="Running Demucs (this may take a minute)...")
52
-
53
- process = await asyncio.create_subprocess_shell(
54
- command,
55
- stdout=asyncio.subprocess.PIPE,
56
- stderr=asyncio.subprocess.PIPE)
57
-
58
- stdout, stderr = await process.communicate()
59
-
60
- if process.returncode != 0:
61
- raise gr.Error(f"Demucs failed to run. Error: {stderr.decode()[:500]}")
62
-
63
- progress(0.8, desc="Locating separated stem files...")
64
- stable_filename_base = os.path.basename(stable_input_path).rsplit('.', 1)[0]
65
- model_folder_name = next(os.walk(output_dir))[1][0]
66
- stems_path = os.path.join(output_dir, model_folder_name, stable_filename_base)
67
-
68
- if not os.path.exists(stems_path):
69
- raise gr.Error(f"Demucs finished, but the output directory was not found!")
70
-
71
- vocals_path = os.path.join(stems_path, "vocals.wav") if os.path.exists(os.path.join(stems_path, "vocals.wav")) else None
72
- drums_path = os.path.join(stems_path, "drums.wav") if os.path.exists(os.path.join(stems_path, "drums.wav")) else None
73
- bass_path = os.path.join(stems_path, "bass.wav") if os.path.exists(os.path.join(stems_path, "bass.wav")) else None
74
- other_filename = "no_vocals.wav" if "2 Stems" in stem_choice else "other.wav"
75
- other_path = os.path.join(stems_path, other_filename) if os.path.exists(os.path.join(stems_path, other_filename)) else None
76
-
77
  os.remove(stable_input_path)
78
  return vocals_path, drums_path, bass_path, other_path
79
  except Exception as e:
80
- print(f"An error occurred: {e}")
81
- raise gr.Error(str(e))
82
 
83
- def slice_stem_real(stem_audio_data, loop_choice, sensitivity, progress=gr.Progress(track_tqdm=True)):
84
  if stem_audio_data is None:
85
- gr.Warning("This stem is empty. Cannot slice.")
86
- return None, None
87
-
88
- sample_rate, y_int = stem_audio_data
89
- y = librosa.util.buf_to_float(y_int)
90
- y_mono = librosa.to_mono(y.T) if y.ndim > 1 else y
91
-
92
- progress(0.1, desc="Detecting BPM...")
93
- tempo, beats = librosa.beat.beat_track(y=y_mono, sr=sample_rate)
94
- bpm = 120 if tempo is None else np.round(tempo)
95
- bpm_int = int(bpm.item())
96
-
97
- if bpm_int == 0:
98
- bpm_int = 120
99
- gr.Warning("BPM detection failed, defaulting to 120 BPM.")
100
- print(f"Detected BPM: {bpm_int}")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
101
 
102
- output_files = []
103
- loops_dir = tempfile.mkdtemp()
104
-
105
- if "One-Shots" in loop_choice:
106
- progress(0.3, desc="Finding transients...")
107
- onset_frames = librosa.onset.onset_detect(y=y_mono, sr=sample_rate, delta=sensitivity, wait=1, pre_avg=1, post_avg=1, post_max=1)
108
- onset_samples = librosa.frames_to_samples(onset_frames)
109
-
110
- progress(0.5, desc="Slicing one-shots...")
111
- if len(onset_samples) > 0:
112
- for i, start_sample in enumerate(onset_samples):
113
- end_sample = onset_samples[i+1] if i+1 < len(onset_samples) else len(y)
114
- slice_data = y[start_sample:end_sample]
115
- filename = os.path.join(loops_dir, f"one_shot_{i+1:03d}.wav")
116
- sf.write(filename, slice_data, sample_rate, subtype='PCM_16')
117
- output_files.append(filename)
118
- if len(onset_samples) > 1:
119
- progress(0.5 + (i / (len(onset_samples) - 1) * 0.5), desc=f"Exporting slice {i+1}...")
120
- else: # Handle bar loops
121
- bars = int(loop_choice.split(" ")[0])
122
- seconds_per_beat = 60.0 / bpm_int
123
- seconds_per_bar = seconds_per_beat * 4
124
- loop_duration_seconds = seconds_per_bar * bars
125
- loop_duration_samples = int(loop_duration_seconds * sample_rate)
126
-
127
- progress(0.4, desc=f"Slicing into {bars}-bar loops...")
128
- num_loops = len(y) // loop_duration_samples
129
- if num_loops == 0:
130
- gr.Warning(f"Audio is too short to create a {bars}-bar loop at {bpm_int} BPM.")
131
- return None, None
132
- for i in range(num_loops):
133
- start_sample = i * loop_duration_samples
134
- end_sample = start_sample + loop_duration_samples
135
- slice_data = y[start_sample:end_sample]
136
- filename = os.path.join(loops_dir, f"{bars}bar_loop_{i+1:03d}_{bpm_int}bpm.wav")
137
- sf.write(filename, slice_data, sample_rate, subtype='PCM_16')
138
- output_files.append(filename)
139
- if num_loops > 1:
140
- progress(0.4 + (i / (num_loops - 1) * 0.6), desc=f"Exporting loop {i+1}...")
141
-
142
- if not output_files:
143
- gr.Warning("No loops or one-shots could be generated.")
144
- return None, None
145
-
146
- return output_files, loops_dir
147
-
148
- async def slice_all_and_zip_real(vocals, drums, bass, other, loop_choice, sensitivity, progress=gr.Progress(track_tqdm=True)):
149
- progress(0, desc="Starting batch slice...")
150
- await asyncio.sleep(0.1)
151
- stems_to_process = {"vocals": vocals, "drums": drums, "bass": bass, "other": other}
152
-
153
- zip_path = "Loop_Architect_Pack.zip"
154
-
155
- num_stems = sum(1 for data in stems_to_process.values() if data is not None)
156
- if num_stems == 0: raise gr.Error("No stems to process! Please separate stems first.")
157
 
158
  with zipfile.ZipFile(zip_path, 'w') as zf:
159
- processed_count = 0
160
- for name, data in stems_to_process.items():
161
- if data is not None:
162
- def update_main_progress(p, desc=""):
163
- overall_progress = (processed_count + p) / num_stems
164
- progress(overall_progress, desc=f"Slicing {name}: {desc}")
165
-
166
- sliced_files, temp_dir = slice_stem_real((data[0], data[1]), loop_choice, sensitivity, progress=update_main_progress)
167
-
168
- if sliced_files:
169
- for loop_file in sliced_files:
170
- arcname = os.path.join(name, os.path.basename(loop_file))
171
- zf.write(loop_file, arcname)
172
-
173
- if temp_dir and os.path.exists(temp_dir):
174
- shutil.rmtree(temp_dir)
175
- processed_count += 1
176
-
177
  progress(1, desc="Pack Ready!")
178
  return zip_path, gr.update(visible=True)
179
 
180
-
181
  # --- Create the full Gradio Interface ---
182
  with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="red")) as demo:
183
  gr.Markdown("# 🎵 Loop Architect")
 
 
 
 
 
 
184
  with gr.Row():
185
  with gr.Column(scale=1):
186
- gr.Markdown("### 1. Separate Stems")
187
- audio_input = gr.Audio(type="filepath", label="Upload a Track")
188
  stem_options = gr.Radio(["4 Stems (Vocals, Drums, Bass, Other)", "2 Stems (Vocals + Instrumental)"], label="Separation Type", value="4 Stems (Vocals, Drums, Bass, Other)")
189
  submit_button = gr.Button("Separate Stems")
190
  with gr.Accordion("Slicing Options", open=True):
191
- loop_options_radio = gr.Radio(["One-Shots (All Transients)", "4 Bar Loops", "8 Bar Loops"], label="Loop Length", value="One-Shots (All Transients)")
 
192
  sensitivity_slider = gr.Slider(minimum=0.01, maximum=0.5, value=0.05, step=0.01, label="One-Shot Sensitivity")
193
- gr.Markdown("### 3. Create Pack")
194
- slice_all_button = gr.Button("Slice All Stems & Create Pack", variant="primary")
195
- download_zip_file = gr.File(label="Download Your Loop Pack", visible=False)
196
  with gr.Column(scale=2):
197
  with gr.Accordion("Separated Stems", open=True):
198
- with gr.Row():
199
- vocals_output = gr.Audio(label="Vocals", scale=4)
200
- slice_vocals_btn = gr.Button("Slice", scale=1)
201
- with gr.Row():
202
- drums_output = gr.Audio(label="Drums", scale=4)
203
- slice_drums_btn = gr.Button("Slice", scale=1)
204
- with gr.Row():
205
- bass_output = gr.Audio(label="Bass", scale=4)
206
- slice_bass_btn = gr.Button("Slice", scale=1)
207
- with gr.Row():
208
- other_output = gr.Audio(label="Other / Instrumental", scale=4)
209
- slice_other_btn = gr.Button("Slice", scale=1)
210
- gr.Markdown("### Sliced Loops / Samples")
211
- loop_gallery = gr.Gallery(label="Generated Loops", columns=8, object_fit="contain", height="auto")
 
 
 
212
 
213
  # --- Define Event Listeners ---
214
- def slice_and_display(stem_data, loop_choice, sensitivity):
215
- files, _ = slice_stem_real(stem_data, loop_choice, sensitivity)
216
- return files
217
-
218
  submit_button.click(fn=separate_stems, inputs=[audio_input, stem_options], outputs=[vocals_output, drums_output, bass_output, other_output])
219
- stem_options.change(fn=update_output_visibility, inputs=[stem_options], outputs=[vocals_output, drums_output, bass_output, other_output])
220
- slice_vocals_btn.click(fn=slice_and_display, inputs=[vocals_output, loop_options_radio, sensitivity_slider], outputs=loop_gallery)
221
- slice_drums_btn.click(fn=slice_and_display, inputs=[drums_output, loop_options_radio, sensitivity_slider], outputs=loop_gallery)
222
- slice_bass_btn.click(fn=slice_and_display, inputs=[bass_output, loop_options_radio, sensitivity_slider], outputs=loop_gallery)
223
- slice_other_btn.click(fn=slice_and_display, inputs=[other_output, loop_options_radio, sensitivity_slider], outputs=loop_gallery)
224
- slice_all_button.click(fn=slice_all_and_zip_real, inputs=[vocals_output, drums_output, bass_output, other_output, loop_options_radio, sensitivity_slider], outputs=[download_zip_file, download_zip_file])
 
 
 
 
 
 
 
225
 
226
  # --- Launch the UI ---
227
- # Use demo.launch() for Hugging Face Spaces
228
- # Use demo.launch(debug=True) for local testing
229
- demo.launch()
 
1
  # 1. Install all necessary libraries for the full application
2
+ !pip install gradio "demucs>=4.0.0" librosa soundfile matplotlib
3
 
4
  # 2. Import libraries
5
  import gradio as gr
 
21
 
22
  def update_output_visibility(choice):
23
  if "2 Stems" in choice:
24
+ return { vocals_output: gr.update(visible=True), drums_output: gr.update(visible=False), bass_output: gr.update(visible=False), other_output: gr.update(visible=True, label="Instrumental (No Vocals)") }
 
 
 
 
 
25
  elif "4 Stems" in choice:
26
+ return { vocals_output: gr.update(visible=True), drums_output: gr.update(visible=True), bass_output: gr.update(visible=True), other_output: gr.update(visible=True, label="Other") }
 
 
 
 
 
27
 
28
  async def separate_stems(audio_file_path, stem_choice, progress=gr.Progress(track_tqdm=True)):
29
  if audio_file_path is None: raise gr.Error("No audio file uploaded!")
30
+ progress(0, desc="Starting..."); await asyncio.sleep(0.1)
31
  try:
32
+ progress(0.05, desc="Preparing audio file..."); original_filename_base = os.path.basename(audio_file_path).rsplit('.', 1)[0]; stable_input_path = f"stable_input_{original_filename_base}.wav"; shutil.copy(audio_file_path, stable_input_path)
33
+ model_arg = "--two-stems=vocals" if "2 Stems" in stem_choice else ""; output_dir = "separated"
 
 
 
 
 
34
  if os.path.exists(output_dir): shutil.rmtree(output_dir)
 
35
  command = f"python3 -m demucs {model_arg} -o \"{output_dir}\" \"{stable_input_path}\""
36
+ progress(0.2, desc="Running Demucs..."); process = await asyncio.create_subprocess_shell(command, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE); await process.communicate()
37
+ if process.returncode != 0: raise gr.Error(f"Demucs failed to run.")
38
+ progress(0.8, desc="Locating separated stem files..."); stable_filename_base = os.path.basename(stable_input_path).rsplit('.', 1)[0]; model_folder_name = next(os.walk(output_dir))[1][0]; stems_path = os.path.join(output_dir, model_folder_name, stable_filename_base)
39
+ if not os.path.exists(stems_path): raise gr.Error(f"Demucs finished, but the output directory was not found!")
40
+ vocals_path = os.path.join(stems_path, "vocals.wav") if os.path.exists(os.path.join(stems_path, "vocals.wav")) else None; drums_path = os.path.join(stems_path, "drums.wav") if os.path.exists(os.path.join(stems_path, "drums.wav")) else None; bass_path = os.path.join(stems_path, "bass.wav") if os.path.exists(os.path.join(stems_path, "bass.wav")) else None
41
+ other_filename = "no_vocals.wav" if "2 Stems" in stem_choice else "other.wav"; other_path = os.path.join(stems_path, other_filename) if os.path.exists(os.path.join(stems_path, other_filename)) else None
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  os.remove(stable_input_path)
43
  return vocals_path, drums_path, bass_path, other_path
44
  except Exception as e:
45
+ print(f"An error occurred: {e}"); raise gr.Error(str(e))
 
46
 
47
+ def visualize_slices(stem_audio_data, progress=gr.Progress(track_tqdm=True)):
48
  if stem_audio_data is None:
49
+ gr.Warning("This stem is empty. Cannot visualize."); return None, None, None
50
+ sample_rate, y_int = stem_audio_data; y = librosa.util.buf_to_float(y_int);
51
+ progress(0.3, desc="Finding transients..."); onset_frames = librosa.onset.onset_detect(y=librosa.to_mono(y.T) if y.ndim > 1 else y, sr=sample_rate, wait=1, pre_avg=1, post_avg=1, post_max=1, delta=0.05)
52
+ onset_times = librosa.frames_to_time(onset_frames, sr=sample_rate)
53
+ progress(0.7, desc="Generating waveform plot...")
54
+ fig, ax = plt.subplots(figsize=(10, 3)); fig.patch.set_facecolor('#1f2937'); ax.set_facecolor('#111827')
55
+ librosa.display.waveshow(y, sr=sample_rate, ax=ax, color='#32f6ff', alpha=0.7)
56
+ for t in onset_times:
57
+ ax.axvline(x=t, color='#ff3b3b', linestyle='--', linewidth=1)
58
+ ax.tick_params(colors='gray'); ax.xaxis.label.set_color('gray'); ax.yaxis.label.set_color('gray'); ax.set_xlabel("Time (s)"); ax.set_ylabel("Amplitude"); ax.set_title("Detected Slices", color='white'); plt.tight_layout()
59
+ progress(1, desc="Done!")
60
+ return fig, onset_times, stem_audio_data
61
+
62
+ def preview_slice(active_stem_audio, onset_times, evt: gr.SelectData):
63
+ if active_stem_audio is None or onset_times is None:
64
+ return None
65
+ sample_rate, y = active_stem_audio; clicked_time = evt.index[0]
66
+ start_time = 0; end_time = len(y) / sample_rate
67
+ for i, t in enumerate(onset_times):
68
+ if t > clicked_time:
69
+ end_time = t; break
70
+ start_time = t
71
+ start_sample = librosa.time_to_samples(start_time, sr=sample_rate); end_sample = librosa.time_to_samples(end_time, sr=sample_rate)
72
+ sliced_audio = y[start_sample:end_sample]
73
+ return (sample_rate, sliced_audio)
74
+
75
+ # --- NEW FUNCTIONS FOR MANAGING THE PACK ---
76
+ def add_slice_to_pack(current_preview, selection_list):
77
+ if current_preview is None:
78
+ gr.Warning("No slice is being previewed to add.")
79
+ return selection_list
80
+ selection_list.append(current_preview)
81
+ gr.Info(f"Slice added! You now have {len(selection_list)} slices in your pack.")
82
+ return selection_list
83
+
84
+ def clear_selection():
85
+ gr.Info("Selection cleared.")
86
+ return []
87
+
88
+ def create_final_pack(selection_list, progress=gr.Progress(track_tqdm=True)):
89
+ if not selection_list:
90
+ raise gr.Error("No slices have been selected to create a pack!")
91
 
92
+ progress(0, desc="Preparing final pack...")
93
+ zip_path = "Custom_Loop_Pack.zip"
94
+ temp_dir = tempfile.mkdtemp()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
95
 
96
  with zipfile.ZipFile(zip_path, 'w') as zf:
97
+ for i, audio_data in enumerate(selection_list):
98
+ progress(i / len(selection_list), desc=f"Adding slice {i+1} to pack...")
99
+ sample_rate, y = audio_data
100
+ filename = os.path.join(temp_dir, f"slice_{i+1:03d}.wav")
101
+ sf.write(filename, y, sample_rate, subtype='PCM_16')
102
+ zf.write(filename, os.path.basename(filename))
103
+
104
+ shutil.rmtree(temp_dir)
 
 
 
 
 
 
 
 
 
 
105
  progress(1, desc="Pack Ready!")
106
  return zip_path, gr.update(visible=True)
107
 
 
108
  # --- Create the full Gradio Interface ---
109
  with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="red")) as demo:
110
  gr.Markdown("# 🎵 Loop Architect")
111
+
112
+ # State components
113
+ onset_times_state = gr.State(value=None)
114
+ active_stem_state = gr.State(value=None)
115
+ selected_slices_state = gr.State(value=[])
116
+
117
  with gr.Row():
118
  with gr.Column(scale=1):
119
+ gr.Markdown("### 1. Separate Stems"); audio_input = gr.Audio(type="filepath", label="Upload a Track")
 
120
  stem_options = gr.Radio(["4 Stems (Vocals, Drums, Bass, Other)", "2 Stems (Vocals + Instrumental)"], label="Separation Type", value="4 Stems (Vocals, Drums, Bass, Other)")
121
  submit_button = gr.Button("Separate Stems")
122
  with gr.Accordion("Slicing Options", open=True):
123
+ gr.Markdown("These options are for the final pack creation.")
124
+ loop_options_radio = gr.Radio(["One-Shots (All Transients)", "4 Bar Loops", "8 Bar Loops"], label="Loop Type (for 'Slice All')", value="One-Shots (All Transients)")
125
  sensitivity_slider = gr.Slider(minimum=0.01, maximum=0.5, value=0.05, step=0.01, label="One-Shot Sensitivity")
126
+
 
 
127
  with gr.Column(scale=2):
128
  with gr.Accordion("Separated Stems", open=True):
129
+ with gr.Row(): vocals_output = gr.Audio(label="Vocals", scale=4); slice_vocals_btn = gr.Button("Visualize Slices", scale=1)
130
+ with gr.Row(): drums_output = gr.Audio(label="Drums", scale=4); slice_drums_btn = gr.Button("Visualize Slices", scale=1)
131
+ with gr.Row(): bass_output = gr.Audio(label="Bass", scale=4); slice_bass_btn = gr.Button("Visualize Slices", scale=1)
132
+ with gr.Row(): other_output = gr.Audio(label="Other / Instrumental", scale=4); slice_other_btn = gr.Button("Visualize Slices", scale=1)
133
+
134
+ gr.Markdown("### Slice Editor")
135
+ slice_plot = gr.Plot(label="Click a region on the waveform to preview a slice")
136
+ with gr.Row():
137
+ preview_player = gr.Audio(label="Slice Preview", scale=3)
138
+ add_to_pack_btn = gr.Button("Add to Pack", variant="primary", scale=1)
139
+
140
+ gr.Markdown("### Your Custom Pack")
141
+ with gr.Row():
142
+ create_pack_btn = gr.Button("Create Pack from Selection", variant="primary")
143
+ clear_selection_btn = gr.Button("Clear Selection")
144
+ selected_gallery = gr.Gallery(label="Selected Slices", columns=8, object_fit="contain", height="auto")
145
+ download_zip_file = gr.File(label="Download Your Custom Pack", visible=False)
146
 
147
  # --- Define Event Listeners ---
 
 
 
 
148
  submit_button.click(fn=separate_stems, inputs=[audio_input, stem_options], outputs=[vocals_output, drums_output, bass_output, other_output])
149
+ stem_options.change(fn=update_output_visibility, inputs=stem_options, outputs=[vocals_output, drums_output, bass_output, other_output])
150
+
151
+ slice_vocals_btn.click(fn=visualize_slices, inputs=vocals_output, outputs=[slice_plot, onset_times_state, active_stem_state])
152
+ slice_drums_btn.click(fn=visualize_slices, inputs=drums_output, outputs=[slice_plot, onset_times_state, active_stem_state])
153
+ slice_bass_btn.click(fn=visualize_slices, inputs=bass_output, outputs=[slice_plot, onset_times_state, active_stem_state])
154
+ slice_other_btn.click(fn=visualize_slices, inputs=other_output, outputs=[slice_plot, onset_times_state, active_stem_state])
155
+
156
+ slice_plot.select(fn=preview_slice, inputs=[active_stem_state, onset_times_state], outputs=preview_player)
157
+
158
+ add_to_pack_btn.click(fn=add_slice_to_pack, inputs=[preview_player, selected_slices_state], outputs=selected_slices_state)
159
+ selected_slices_state.change(fn=lambda x: x, inputs=selected_slices_state, outputs=selected_gallery) # Update gallery when state changes
160
+ clear_selection_btn.click(fn=clear_selection, outputs=selected_slices_state)
161
+ create_pack_btn.click(fn=create_final_pack, inputs=selected_slices_state, outputs=[download_zip_file, download_zip_file])
162
 
163
  # --- Launch the UI ---
164
+ demo.launch(debug=True)