| --- |
| language: |
| - en |
| license: apache-2.0 |
| size_categories: |
| - 10M<n<100M |
| task_categories: |
| - audio-to-audio |
| pretty_name: Hive Dataset |
| arxiv: 2601.22599 |
| tags: |
| - audio |
| - sound-separation |
| - universal-sound-separation |
| - audio-mixing |
| - audioset |
| dataset_info: |
| features: |
| - name: mix_id |
| dtype: string |
| - name: split |
| dtype: string |
| - name: sample_rate |
| dtype: int32 |
| - name: target_duration |
| dtype: float64 |
| - name: num_sources |
| dtype: int32 |
| - name: sources |
| sequence: |
| - name: source_id |
| dtype: string |
| - name: path |
| dtype: string |
| - name: label |
| dtype: string |
| - name: crop_start_second |
| dtype: float64 |
| - name: crop_end_second |
| dtype: float64 |
| - name: chunk_start_second |
| dtype: float64 |
| - name: chunk_end_second |
| dtype: float64 |
| - name: rms_gain |
| dtype: float64 |
| - name: snr_db |
| dtype: float64 |
| - name: applied_weight |
| dtype: float64 |
| - name: global_normalization_factor |
| dtype: float64 |
| - name: final_max_amplitude |
| dtype: float64 |
| splits: |
| - name: train |
| num_examples: 5000000 |
| - name: validation |
| num_examples: 500000 |
| - name: test |
| num_examples: 100000 |
| --- |
| |
| <h1 align="center">A Semantically Consistent Dataset for Data-Efficient Query-Based Universal Sound Separation</h1> |
| <p align="center"> |
| <img src="asserts/logo.png" alt="Logo" width="250"/> |
| </p> |
| <p align="center"> |
| <strong>Kai Li<sup>*</sup>, Jintao Cheng<sup>*</sup>, Chang Zeng, Zijun Yan, Helin Wang, Zixiong Su, Bo Zheng, Xiaolin Hu</strong><br> |
| <strong>Tsinghua University, Shanda AI, Johns Hopkins University</strong><br> |
| <strong><sup>*</sup>Equal contribution</strong><br> |
| <strong>Completed during Kai Li's internship at Shanda AI.</strong><br> |
| <a href="https://arxiv.org/abs/2601.22599">📜 Arxiv 2026</a> | <a href="https://github.com/ShandaAI/Hive">💻 Code</a> | <a href="https://shandaai.github.io/Hive/">🎶 Demo</a> |
| </p> |
| |
| ## Usage |
|
|
| ```python |
| from datasets import load_dataset |
| |
| # Load full dataset |
| dataset = load_dataset("ShandaAI/Hive") |
| |
| # Load specific split |
| train_data = load_dataset("ShandaAI/Hive", split="train") |
| |
| # Streaming mode (recommended for large datasets) |
| dataset = load_dataset("ShandaAI/Hive", streaming=True) |
| ``` |
|
|
| ## 📄 Dataset Description |
|
|
| **Hive** is a high-quality synthetic dataset designed for **Universal Sound Separation (USS)**. Unlike traditional methods relying on weakly-labeled in-the-wild data, Hive leverages an automated data collection pipeline to mine high-purity single-event segments from complex acoustic environments and synthesizes mixtures with semantically consistent constraints. |
|
|
| ### Key Features |
|
|
| - **Purity over Scale**: 2.4k hours achieving competitive performance with million-hour baselines (~0.2% data scale) |
| - **Single-label Clean Supervision**: Rigorous semantic-acoustic alignment eliminating co-occurrence noise |
| - **Semantically Consistent Mixing**: Logic-based co-occurrence matrix ensuring realistic acoustic scenes |
| - **High Fidelity**: 44.1kHz sample rate for high-quality audio |
|
|
| ### Dataset Scale |
|
|
| | Metric | Value | |
| |--------|-------| |
| | **Training Set Raw Audio** | 2,442 hours | |
| | **Val & Test Set Raw Audio** | 292 hours | |
| | **Mixed Samples** | 19.6M mixtures | |
| | **Total Mixed Duration** | ~22.4k hours | |
| | **Label Categories** | 283 classes | |
| | **Sample Rate** | 44.1 kHz | |
| | **Training Sample Duration** | 4 seconds | |
| | **Test Sample Duration** | 10 seconds | |
|
|
| ### Dataset Splits |
|
|
| | Split | Samples | Description | |
| |-------|---------|-------------| |
| | Train | 17.5M | Training mixtures (4s duration) | |
| | Validation | 1.75M | Validation mixtures | |
| | Test | 350k | Test mixtures (10s duration) | |
|
|
| --- |
|
|
| ## 📂 Dataset Structure |
|
|
| ### Directory Organization |
|
|
| ``` |
| hive-datasets-parquet/ |
| ├── README.md |
| ├── train/ |
| │ └── data.parquet |
| ├── validation/ |
| │ └── data.parquet |
| └── test/ |
| └── data.parquet |
| ``` |
|
|
| Each split contains a single Parquet file with all mixture metadata. The `num_sources` field indicates the number of sources (2-5) for each mixture. |
|
|
| --- |
|
|
| ## 📋 Data Fields |
|
|
| ### JSON Schema |
|
|
| Each JSON object contains complete generation parameters for reproducing a mixture sample: |
|
|
| ```python |
| { |
| "mix_id": "sample_00000003", |
| "split": "train", |
| "sample_rate": 44100, |
| "target_duration": 4.0, |
| "num_sources": 2, |
| "sources": { |
| "source_id": ["s1", "s2"], |
| "path": ["relative/path/to/audio1", "relative/path/to/audio2"], |
| "label": ["Ocean", "Rain"], |
| "crop_start_second": [1.396, 2.5], |
| "crop_end_second": [5.396, 6.5], |
| "chunk_start_second": [35.0, 20.0], |
| "chunk_end_second": [45.0, 30.0], |
| "rms_gain": [3.546, 2.1], |
| "snr_db": [0.0, -3.0], |
| "applied_weight": [3.546, 1.487] |
| }, |
| "global_normalization_factor": 0.786, |
| "final_max_amplitude": 0.95 |
| } |
| ``` |
|
|
| ### Field Descriptions |
|
|
| #### 1. Basic Info Fields |
|
|
| | Field | Type | Description | |
| |-------|------|-------------| |
| | `mix_id` | string | Unique identifier for the mixture task | |
| | `split` | string | Dataset partition (`train` / `validation` / `test`) | |
| | `sample_rate` | int32 | Audio sample rate in Hz (44100) | |
| | `target_duration` | float64 | Target duration in seconds (4.0 for train, 10.0 for test) | |
| | `num_sources` | int32 | Number of audio sources in this mixture (2-5) | |
|
|
| #### 2. Source Information (`sources`) |
|
|
| Metadata required to reproduce the mixing process for each audio source. Stored in columnar format (dict of lists) for efficient Parquet storage: |
|
|
| | Field | Type | Description | |
| |-------|------|-------------| |
| | `source_id` | list[string] | Source identifiers (`s1`, `s2`, ...) | |
| | `path` | list[string] | Relative paths to the source audio files | |
| | `label` | list[string] | AudioSet ontology labels for each source | |
| | `chunk_start_second` | list[float64] | Start times (seconds) for reading from original audio files | |
| | `chunk_end_second` | list[float64] | End times (seconds) for reading from original audio files | |
| | `crop_start_second` | list[float64] | Precise start positions (seconds) for reproducible random extraction | |
| | `crop_end_second` | list[float64] | Precise end positions (seconds) for reproducible random extraction | |
| | `rms_gain` | list[float64] | Energy normalization coefficients: $\text{target\_rms} / \text{current\_rms}$ | |
| | `snr_db` | list[float64] | Signal-to-noise ratios in dB assigned to each source | |
| | `applied_weight` | list[float64] | Final scaling weights: $\text{rms\_gain} \times 10^{(\text{snr\_db} / 20)}$ | |
|
|
| #### 3. Mixing Parameters |
|
|
| Global processing parameters after combining multiple audio sources: |
|
|
| | Field | Type | Description | |
| |-------|------|-------------| |
| | `global_normalization_factor` | float64 | Anti-clipping scaling coefficient: $0.95 / \text{max\_val}$ | |
| | `final_max_amplitude` | float64 | Maximum amplitude threshold (0.95) to prevent bit-depth overflow | |
| |
| ### Detailed Field Explanations |
| |
| #### Cropping Logic |
| - `chunk_start/end_second`: Defines the reading interval from the original audio file |
| - `crop_start/end_second`: Records the precise random cropping position, ensuring exact reproducibility across runs |
| |
| #### Energy Normalization (`rms_gain`) |
| Adjusts different audio sources to the same energy level: |
| $$\text{rms\_gain} = \frac{\text{target\_rms}}{\text{current\_rms}}$$ |
| |
| #### Signal-to-Noise Ratio (`snr_db`) |
| The SNR value assigned to each source, sampled from a predefined range using `random.uniform(snr_range[0], snr_range[1])`. |
|
|
| #### Applied Weight |
| The comprehensive scaling weight combining energy normalization and SNR adjustment: |
| $$\text{applied\_weight} = \text{rms\_gain} \times 10^{(\text{snr\_db} / 20)}$$ |
| |
| This is the final coefficient applied to the original waveform. |
| |
| #### Global Normalization Factor |
| Prevents audio clipping after mixing: |
| $$\text{global\_normalization\_factor} = \frac{0.95}{\text{max\_val}}$$ |
|
|
| Where `max_val` is the **peak amplitude (absolute value)** of the mixed signal. |
|
|
| --- |
|
|
| ## 🔧 Usage |
|
|
| ### Download Metadata |
|
|
| ```python |
| from datasets import load_dataset |
| |
| # Load specific split and mixture type |
| dataset = load_dataset("ShandaAI/Hive", split="train") |
| ``` |
|
|
| ### Generate Mixed Audio |
|
|
| Please refer to the [official GitHub repository](https://github.com/ShandaAI/Hive) for the complete audio generation pipeline. |
|
|
| ```bash |
| # Clone the repository |
| git clone https://github.com/ShandaAI/Hive.git |
| cd Hive/hive_dataset |
| |
| # Generate mixtures from metadata |
| python mix_from_metadata/mix_from_metadata.py \ |
| --metadata_dir /path/to/downloaded/metadata \ |
| --output_dir ./hive_dataset \ |
| --dataset_paths dataset_paths.json \ |
| --num_processes 16 |
| ``` |
|
|
| --- |
|
|
| ## 📚 Source Datasets |
|
|
| Hive integrates **12 public datasets** to construct a long-tailed acoustic space: |
|
|
| | # | Dataset | Clips | Duration (h) | License | |
| |---|---------|-------|--------------|---------| |
| | 1 | BBC Sound Effects | 369,603 | 1,020.62 | Remix License | |
| | 2 | AudioSet | 326,890 | 896.61 | CC BY | |
| | 3 | VGGSound | 115,191 | 319.10 | CC BY 4.0 | |
| | 4 | MUSIC21 | 32,701 | 90.28 | YouTube Standard | |
| | 5 | FreeSound | 17,451 | 46.90 | CC0/BY/BY-NC | |
| | 6 | ClothoV2 | 14,759 | 38.19 | Non-Commercial Research | |
| | 7 | Voicebank-DEMAND | 12,376 | 9.94 | CC BY 4.0 | |
| | 8 | AVE | 3,054 | 6.91 | CC BY-NC-SA | |
| | 9 | SoundBible | 2,501 | 5.78 | CC BY 4.0 | |
| | 10 | DCASE | 1,969 | 5.46 | Academic Use | |
| | 11 | ESC50 | 1,433 | 1.99 | CC BY-NC 3.0 | |
| | 12 | FSD50K | 636 | 0.80 | Creative Commons | |
| | | **Total** | **898,564** | **2,442.60** | | |
|
|
| **Important Note**: This repository releases only **metadata** (JSON files containing mixing parameters and source references) for reproducibility. Users must independently download and prepare the source datasets according to their respective licenses. |
|
|
| --- |
|
|
| ## 📖 Citation |
|
|
| If you use this dataset, please cite: |
|
|
| ```bibtex |
| @article{li2026hive, |
| title={A Semantically Consistent Dataset for Data-Efficient Query-Based Universal Sound Separation}, |
| author={Li, Kai and Cheng, Jintao and Zeng, Chang and Yan, Zijun and Wang, Helin and Su, Zixiong and Zheng, Bo and Hu, Xiaolin}, |
| journal={arXiv preprint arXiv:2601.22599}, |
| year={2026} |
| } |
| ``` |
|
|
| --- |
|
|
| ## ⚖️ License |
|
|
| This dataset metadata is released under the **Apache License 2.0**. |
|
|
| Please note that the source audio files are subject to their original licenses. Users must comply with the respective licenses when using the source datasets. |
|
|
| --- |
|
|
| ## 🙏 Acknowledgments |
|
|
| We extend our gratitude to the researchers and organizations who curated the foundational datasets that made Hive possible: |
|
|
| - **BBC Sound Effects** - Professional-grade recordings with broadcast-level fidelity |
| - **AudioSet** (Google) - Large-scale audio benchmark |
| - **VGGSound** (University of Oxford) - Real-world acoustic diversity |
| - **FreeSound** (MTG-UPF) - Rich crowdsourced soundscapes |
| - And all other contributing datasets |
|
|
| --- |
|
|
| ## 📬 Contact |
|
|
| For questions or issues, please open an issue on the [GitHub repository](https://github.com/ShandaAI/Hive) or contact the authors. |