| | --- |
| | license: cc-by-nc-sa-4.0 |
| | task_categories: |
| | - audio-to-audio |
| | language: |
| | - en |
| | tags: |
| | - SFZ |
| | - SFZ-Instruments |
| | - Instruments |
| | - MIDI |
| | - MIDI-Instruments |
| | pretty_name: sfzinstruments |
| | size_categories: |
| | - n<1K |
| | --- |
| | |
| | # SFZ Instruments |
| | ## Select open source musical instruments which use the SFZ format |
| |
|
| | *** |
| | |
| | ## Attribution |
| | |
| | ### All SFZ instruments were sourced from [sfzinstruments](https://github.com/sfzinstruments) |
| | |
| | *** |
| |
|
| | ## How to use (Linux/Ubuntu) |
| |
|
| | ### 1) Unzip the dataset into some folder |
| |
|
| | ### 2) Setup environment |
| |
|
| | ```sh |
| | !sudo apt install libsndfile1 ffmpeg libjack-jackd2-dev -y |
| | !pip install soundfile numpy scipy tqdm |
| | ``` |
| |
|
| | ### 3) Build and install [sfizz](https://github.com/sfztools/sfizz) |
| |
|
| | ```sh |
| | !sudo apt install cmake g++ git libsndfile1-dev libjack-jackd2-dev \ |
| | libsamplerate0-dev libboost-dev libzstd-dev \ |
| | libcurl4-openssl-dev libx11-dev -y |
| | |
| | !git clone https://github.com/sfztools/sfizz.git |
| | %cd sfizz |
| | !mkdir build |
| | %cd build |
| | !cmake .. -DCMAKE_BUILD_TYPE=Release |
| | !make -j$(nproc) |
| | !sudo make install |
| | !sudo ldconfig |
| | ``` |
| |
|
| | ### 4) Use the following python script to render your MIDIs |
| |
|
| | ```python |
| | #!/usr/bin/env python3 |
| | """ |
| | render_and_mix.py |
| | |
| | Render MIDI+SFZ pairs using sfizz_render, mix them, apply basic mastering (limiter + normalize), |
| | and optionally run ffmpeg loudness normalization. |
| | |
| | Dependencies: |
| | - sfizz_render (system binary) |
| | - libsndfile (system) |
| | - Python packages: soundfile, numpy, scipy, tqdm |
| | - Optional: ffmpeg (for LUFS normalization) |
| | """ |
| | |
| | import os |
| | import shutil |
| | import subprocess |
| | import tempfile |
| | from concurrent.futures import ThreadPoolExecutor, as_completed |
| | from pathlib import Path |
| | from typing import Dict, Optional |
| | |
| | import numpy as np |
| | import soundfile as sf |
| | from scipy.signal import fftconvolve |
| | from tqdm import tqdm |
| | |
| | # ---------- Utility audio functions ---------- |
| | |
| | def db_to_linear(db: float) -> float: |
| | return 10.0 ** (db / 20.0) |
| | |
| | def linear_to_db(x: float) -> float: |
| | return 20.0 * np.log10(np.maximum(x, 1e-12)) |
| | |
| | def apply_pan(stereo: np.ndarray, pan: float) -> np.ndarray: |
| | # pan: -1 (left) .. +1 (right) |
| | left_gain = np.cos((pan + 1) * (np.pi / 4)) |
| | right_gain = np.sin((pan + 1) * (np.pi / 4)) |
| | stereo[:, 0] *= left_gain |
| | stereo[:, 1] *= right_gain |
| | return stereo |
| | |
| | def ensure_stereo(arr: np.ndarray) -> np.ndarray: |
| | if arr.ndim == 1: |
| | return np.stack([arr, arr], axis=1) |
| | if arr.shape[1] == 1: |
| | return np.repeat(arr, 2, axis=1) |
| | return arr[:, :2] |
| | |
| | def soft_limiter(signal: np.ndarray, threshold: float = 0.98, release: float = 0.01, sample_rate: int = 48000) -> np.ndarray: |
| | # Simple per-sample soft clipping with smoothing |
| | out = np.copy(signal) |
| | # apply tanh-style soft clip scaled to threshold |
| | scale = 1.0 / threshold |
| | out = np.tanh(out * scale) / scale |
| | return out |
| | |
| | def normalize_peak(signal: np.ndarray, target_dbfs: float = -1.0) -> np.ndarray: |
| | peak = np.max(np.abs(signal)) |
| | if peak <= 0: |
| | return signal |
| | target_lin = db_to_linear(target_dbfs) |
| | gain = target_lin / peak |
| | return signal * gain |
| | |
| | # ---------- sfizz_render wrapper ---------- |
| | |
| | def find_sfizz_render() -> Optional[str]: |
| | # Try common binary names |
| | for name in ("sfizz_render", "sfizz-render", "sfizz_render.exe"): |
| | path = shutil.which(name) |
| | if path: |
| | return path |
| | return None |
| | |
| | def render_with_sfizz(sfizz_bin: str, midi_path: str, sfz_path: str, out_wav: str, |
| | sample_rate: int = 48000, quality: int = 3, polyphony: int = 256, |
| | use_eot: bool = True, verbose: bool = False) -> None: |
| | cmd = [ |
| | sfizz_bin, |
| | "--midi", str(midi_path), |
| | "--sfz", str(sfz_path), |
| | "--wav", str(out_wav), |
| | "--samplerate", str(sample_rate), |
| | "--quality", str(quality), |
| | "--polyphony", str(polyphony), |
| | ] |
| | if use_eot: |
| | cmd.append("--use-eot") |
| | if verbose: |
| | cmd.append("--verbose") |
| | # Run and raise on error |
| | subprocess.run(cmd, check=True) |
| | |
| | # ---------- Main render and mix function ---------- |
| | |
| | def render_and_mix( |
| | midi_sfz_map: Dict[str, str], |
| | out_path: str, |
| | *, |
| | sample_rate: int = 48000, |
| | quality: int = 3, |
| | polyphony: int = 256, |
| | track_options: Optional[Dict[str, Dict]] = None, |
| | normalize_lufs: Optional[float] = None, |
| | use_eot: bool = True, |
| | workers: int = 2, |
| | verbose: bool = False |
| | ) -> None: |
| | """ |
| | Render each MIDI->SFZ pair, mix, post-process, and write final WAV to out_path. |
| | """ |
| | sfizz_bin = find_sfizz_render() |
| | if not sfizz_bin: |
| | raise FileNotFoundError("sfizz_render binary not found in PATH. Install sfizz-render first.") |
| | |
| | tmpdir = Path(tempfile.mkdtemp(prefix="sfizz_render_")) |
| | rendered_files = {} |
| | |
| | # Render in parallel |
| | with ThreadPoolExecutor(max_workers=workers) as ex: |
| | futures = {} |
| | for midi, sfz in midi_sfz_map.items(): |
| | midi_p = Path(midi) |
| | sfz_p = Path(sfz) |
| | if not midi_p.exists(): |
| | raise FileNotFoundError(f"MIDI file not found: {midi}") |
| | if not sfz_p.exists(): |
| | raise FileNotFoundError(f"SFZ file not found: {sfz}") |
| | out_wav = tmpdir / (midi_p.stem + "_" + sfz_p.stem + ".wav") |
| | futures[ex.submit(render_with_sfizz, sfizz_bin, str(midi_p), str(sfz_p), str(out_wav), |
| | sample_rate, quality, polyphony, use_eot, verbose)] = (midi, str(out_wav)) |
| | |
| | # Wait and collect |
| | for fut in tqdm(as_completed(futures), total=len(futures), desc="Rendering"): |
| | midi_key, wav_path = futures[fut] |
| | fut.result() # will raise if render failed |
| | rendered_files[midi_key] = wav_path |
| | |
| | # Load and align |
| | tracks = [] |
| | max_len = 0 |
| | for midi_key, wav_path in rendered_files.items(): |
| | data, sr = sf.read(wav_path, always_2d=True) |
| | if sr != sample_rate: |
| | # resample if needed (simple linear resample) |
| | import math |
| | ratio = sample_rate / sr |
| | new_len = int(math.ceil(data.shape[0] * ratio)) |
| | # use scipy.signal.resample for decent quality |
| | from scipy.signal import resample |
| | data = resample(data, new_len, axis=0) |
| | data = ensure_stereo(data) |
| | tracks.append((midi_key, data)) |
| | if data.shape[0] > max_len: |
| | max_len = data.shape[0] |
| | |
| | # Prepare final mix buffer |
| | mix = np.zeros((max_len, 2), dtype=np.float32) |
| | |
| | # Apply per-track options and mix |
| | for midi_key, data in tracks: |
| | opts = (track_options or {}).get(midi_key, {}) |
| | gain_db = float(opts.get("gain_db", 0.0)) |
| | pan = float(opts.get("pan", 0.0)) |
| | gain_lin = db_to_linear(gain_db) |
| | # pad to max_len |
| | pad_len = max_len - data.shape[0] |
| | if pad_len > 0: |
| | data = np.vstack([data, np.zeros((pad_len, 2), dtype=data.dtype)]) |
| | data = data.astype(np.float32) * gain_lin |
| | data = apply_pan(data, pan) |
| | mix[:data.shape[0], :] += data |
| | |
| | # Basic safety: prevent NaNs/Infs |
| | mix = np.nan_to_num(mix, nan=0.0, posinf=0.0, neginf=0.0) |
| | |
| | # Apply soft limiter and normalization |
| | mix = soft_limiter(mix, threshold=0.98, sample_rate=sample_rate) |
| | mix = normalize_peak(mix, target_dbfs=-1.0) |
| | |
| | # Write intermediate file |
| | intermediate = tmpdir / "mixed_intermediate.wav" |
| | sf.write(str(intermediate), mix, samplerate=sample_rate, subtype="PCM_24") |
| | |
| | # Optional LUFS normalization via ffmpeg loudnorm |
| | final_out = Path(out_path) |
| | if normalize_lufs is not None: |
| | ffmpeg = shutil.which("ffmpeg") |
| | if not ffmpeg: |
| | raise FileNotFoundError("ffmpeg not found but normalize_lufs requested.") |
| | # two-pass loudnorm recommended; here we do a single-pass approximate target |
| | cmd = [ |
| | ffmpeg, "-y", "-i", str(intermediate), |
| | "-af", f"loudnorm=I={normalize_lufs}:TP=-1.5:LRA=11", |
| | "-ar", str(sample_rate), |
| | "-ac", "2", |
| | "-c:a", "pcm_s24le", |
| | str(final_out) |
| | ] |
| | subprocess.run(cmd, check=True) |
| | else: |
| | # move intermediate to final |
| | shutil.move(str(intermediate), str(final_out)) |
| | |
| | # cleanup |
| | try: |
| | shutil.rmtree(tmpdir) |
| | except Exception: |
| | pass |
| | |
| | # ---------- Example usage ---------- |
| | if __name__ == "__main__": |
| | # Example mapping: two MIDI files each with their SFZ instrument |
| | mapping = { |
| | "midi/drums.mid": "sfz/drumkit.sfz", |
| | "midi/piano.mid": "sfz/grand_piano.sfz", |
| | } |
| | track_opts = { |
| | "midi/drums.mid": {"gain_db": -1.5, "pan": 0.0}, |
| | "midi/piano.mid": {"gain_db": -3.0, "pan": -0.1}, |
| | } |
| | render_and_mix(mapping, "final_mix.wav", sample_rate=48000, quality=3, polyphony=256, |
| | track_options=track_opts, normalize_lufs=-14.0, use_eot=True, workers=2, verbose=False) |
| | |
| | ``` |
| |
|
| | *** |
| | |
| | ### Project Los Angeles |
| | ### Tegridy Code 2026 |