RE-USE / app.py
szuweifu's picture
Upload app.py
aa3114b verified
# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.
#
# NVIDIA CORPORATION and its licensors retain all intellectual property
# and proprietary rights in and to this software, related documentation
# and any modifications thereto. Any use, reproduction, disclosure or
# distribution of this software and related documentation without an express
# license agreement from NVIDIA CORPORATION is strictly prohibited.
import shlex
import subprocess
import spaces
import gradio as gr
def install_mamba():
subprocess.run(shlex.split("pip install https://github.com/state-spaces/mamba/releases/download/v2.2.2/mamba_ssm-2.2.2+cu122torch2.3cxx11abiFALSE-cp310-cp310-linux_x86_64.whl"))
install_mamba()
ABOUT = """
# RE-USE: A universal speech enhancement model for diverse degradations, sampling rates, and languages.
Upload or record a noisy clip, then click **Enhance** to listen to the result and view its spectrogram.
(ref: https://huggingface.co/spaces/rc19477/Speech_Enhancement_Mamba)
"""
import torch
import torchaudio
import torch.nn as nn
import librosa
import matplotlib.pyplot as plt
import numpy as np
from models.stfts import mag_phase_stft, mag_phase_istft
from models.generator_SEMamba_time_d4 import SEMamba
from utils.util import load_config, pad_or_trim_to_match
from huggingface_hub import hf_hub_download
RELU = nn.ReLU()
def make_even(value):
value = int(round(value))
return value if value % 2 == 0 else value + 1
device = "cuda"
cfg1 = load_config('recipes/USEMamba_30x1_lr_00002_norm_05_vq_065_nfft_320_hop_40_NRIR_012_pha_0005_com_04_early_001.yaml')
n_fft, hop_size, win_size = cfg1['stft_cfg']['n_fft'], cfg1['stft_cfg']['hop_size'], cfg1['stft_cfg']['win_size']
compress_factor = cfg1['model_cfg']['compress_factor']
sampling_rate = cfg1['stft_cfg']['sampling_rate']
@spaces.GPU
def enhance(filepath, low_pass_sampling_rate, target_sampling_rate):
USE_model = SEMamba.from_pretrained("nvidia/RE-USE", cfg=cfg1).to(device)
USE_model.eval()
with torch.no_grad():
noisy_wav, noisy_sr = torchaudio.load(filepath)
torchaudio.save("original.wav", noisy_wav.cpu(), noisy_sr)
original_noisy_wav = noisy_wav
original_sr = noisy_sr
if target_sampling_rate != '':
if low_pass_sampling_rate != '':
opts = {"res_type": "kaiser_best"}
noisy_wav = torch.tensor(librosa.resample(noisy_wav.cpu().numpy(), orig_sr=noisy_sr, target_sr=int(low_pass_sampling_rate), **opts))
noisy_sr = int(low_pass_sampling_rate)
opts = {"res_type": "kaiser_best"}
noisy_wav = librosa.resample(noisy_wav.cpu().numpy(), orig_sr=noisy_sr, target_sr=int(target_sampling_rate), **opts)
noisy_sr = int(target_sampling_rate)
noisy_wav = torch.FloatTensor(noisy_wav).to(device)
n_fft_scaled = make_even(n_fft * noisy_sr // sampling_rate)
hop_size_scaled = make_even(hop_size * noisy_sr // sampling_rate)
win_size_scaled = make_even(win_size * noisy_sr // sampling_rate)
noisy_mag, noisy_pha, noisy_com = mag_phase_stft(
noisy_wav,
n_fft=n_fft_scaled,
hop_size=hop_size_scaled,
win_size=win_size_scaled,
compress_factor=compress_factor,
center=True,
addeps=False
)
amp_g, pha_g, _ = USE_model(noisy_mag, noisy_pha)
# To remove "strange sweep artifact"
mag = torch.expm1(RELU(amp_g)) # [1, F, T]
zero_portion = torch.sum(mag==0, 1)/mag.shape[1]
amp_g[:,:,(zero_portion>0.5)[0]] = 0
audio_g = mag_phase_istft(amp_g, pha_g, n_fft_scaled, hop_size_scaled, win_size_scaled, compress_factor)
audio_g = pad_or_trim_to_match(noisy_wav.detach(), audio_g, pad_value=1e-8) # Align lengths using epsilon padding
assert audio_g.shape == noisy_wav.shape, audio_g.shape
# write file
torchaudio.save("enhanced.wav", audio_g.cpu(), noisy_sr)
# spectrograms
fig, axs = plt.subplots(1, 2, figsize=(16, 4))
# noisy
D_noisy = librosa.stft(original_noisy_wav[0].cpu().numpy(), n_fft=512, hop_length=256)
S_noisy = librosa.amplitude_to_db(np.abs(D_noisy), ref=np.max)
librosa.display.specshow(S_noisy, sr=original_sr, hop_length=256, x_axis="time", y_axis="hz", ax=axs[0], vmax=0)
axs[0].set_title("Noisy Spectrogram")
# enhanced
D_clean = librosa.stft(audio_g.cpu().numpy(), n_fft=512, hop_length=256)
S_clean = librosa.amplitude_to_db(np.abs(D_clean), ref=np.max)
librosa.display.specshow(S_clean[0], sr=noisy_sr, hop_length=256, x_axis="time", y_axis="hz", ax=axs[1], vmax=0)
axs[1].set_title("Enhanced Spectrogram")
plt.tight_layout()
return "original.wav", "enhanced.wav", fig
with gr.Blocks() as demo:
gr.Markdown(ABOUT)
gr.Markdown("**Note 1**: For bandwidth extension, the performance may be affected by the characteristics of the input data, particularly the cutoff pattern. A simple solution is to apply low-pass filtering beforehand.")
gr.Markdown("**Note 2**: When processing long input audio, out-of-memory (OOM) errors may occur. To address this, use the chunk-wise inference implementation provided on the Hugging Face.")
with gr.Row():
with gr.Column():
# Create Tabs to separate Audio and Video sessions
with gr.Tabs():
with gr.TabItem("Audio Upload"):
# gr.Audio works great for standard audio files
input_audio = gr.Audio(label="Input Audio", type="filepath")
with gr.TabItem("Video Upload (.mp4, .mov)"):
# gr.File handles .mp4 and .mov without errors
input_video = gr.File(label="Input Video", file_types=[".mp4", ".mov"])
target_sampling_rate = gr.Textbox(label="(Optional) Enter target sampling rate for bandwidth extension:")
low_pass_sampling_rate = gr.Textbox(label="(Optional) Enter target sampling rate for pre-low-pass filtering before bandwidth extension:")
# Helper to unify the input: we use a hidden state to store which one was used
active_input = gr.State()
enhance_btn = gr.Button("Enhance")
with gr.Row():
input_audio_player = gr.Audio(label="Original Input Audio", type="filepath")
output_audio = gr.Audio(label="Enhanced Audio", type="filepath")
plot_output = gr.Plot(label="Spectrograms")
# This function determines which input (audio tab or video tab) to send to your model
def unified_enhance(audio_path, video_path, lp_sr, target_sr):
# Determine which path is valid (the one from the active tab)
# Note: input_video returns a file object, so we get its .name
final_path = audio_path if audio_path else video_path
return enhance(final_path, lp_sr, target_sr)
enhance_btn.click(
fn=unified_enhance,
inputs=[input_audio, input_video, low_pass_sampling_rate, target_sampling_rate],
outputs=[input_audio_player, output_audio, plot_output]
)
demo.queue().launch(share=True)