JusperLee commited on
Commit
bd13b48
ยท
verified ยท
1 Parent(s): 867b3a9

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,3 +1,326 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ task_categories:
4
+ - audio-classification
5
+ - audio-to-audio
6
+ language:
7
+ - en
8
+ tags:
9
+ - audio
10
+ - sound-separation
11
+ - universal-sound-separation
12
+ - audio-mixing
13
+ - audioset
14
+ pretty_name: Hive Dataset
15
+ size_categories:
16
+ - 10M<n<100M
17
+ dataset_info:
18
+ features:
19
+ - name: mix_id
20
+ dtype: string
21
+ - name: split
22
+ dtype: string
23
+ - name: sample_rate
24
+ dtype: int32
25
+ - name: target_duration
26
+ dtype: float64
27
+ - name: num_sources
28
+ dtype: int32
29
+ - name: sources
30
+ sequence:
31
+ - name: source_id
32
+ dtype: string
33
+ - name: path
34
+ dtype: string
35
+ - name: label
36
+ dtype: string
37
+ - name: crop_start_second
38
+ dtype: float64
39
+ - name: crop_end_second
40
+ dtype: float64
41
+ - name: chunk_start_second
42
+ dtype: float64
43
+ - name: chunk_end_second
44
+ dtype: float64
45
+ - name: rms_gain
46
+ dtype: float64
47
+ - name: snr_db
48
+ dtype: float64
49
+ - name: applied_weight
50
+ dtype: float64
51
+ - name: global_normalization_factor
52
+ dtype: float64
53
+ - name: final_max_amplitude
54
+ dtype: float64
55
+ splits:
56
+ - name: train
57
+ num_examples: 5000000
58
+ - name: validation
59
+ num_examples: 500000
60
+ - name: test
61
+ num_examples: 100000
62
+ ---
63
+
64
+ <h1 align="center">A Semantically Consistent Dataset for Data-Efficient Query-Based Universal Sound Separation</h1>
65
+ <p align="center">
66
+ <img src="asserts/logo.png" alt="Logo" width="250"/>
67
+ </p>
68
+ <p align="center">
69
+ <strong>Kai Li<sup>*</sup>, Jintao Cheng<sup>*</sup>, Chang Zeng, Zijun Yan, Helin Wang, Zixiong Su, Bo Zheng, Xiaolin Hu</strong><br>
70
+ <strong>Tsinghua University, Shanda AI, Johns Hopkins University</strong><br>
71
+ <strong><sup>*</sup>Equal contribution</strong><br>
72
+ <strong>Completed during Kai Li's internship at Shanda AI.</strong><br>
73
+ <a href="#">๐Ÿ“œ Arxiv 2026</a> | <a href="#">๐ŸŽถ Demo</a> | <a href="#">๐Ÿค— Dataset</a>
74
+ </p>
75
+
76
+ [![visitors](https://visitor-badge.laobi.icu/badge?page_id=ShandaAI.Hive)](https://huggingface.co/datasets/ShandaAI/Hive) [![GitHub stars](https://img.shields.io/github/stars/ShandaAI/Hive?style=social)](https://github.com/ShandaAI/Hive) [![license](https://img.shields.io/badge/license-Apache%202.0-blue.svg)](LICENSE)
77
+
78
+ ## Usage
79
+
80
+ ```python
81
+ from datasets import load_dataset
82
+
83
+ # Load full dataset
84
+ dataset = load_dataset("ShandaAI/Hive")
85
+
86
+ # Load specific split
87
+ train_data = load_dataset("ShandaAI/Hive", split="train")
88
+
89
+ # Streaming mode (recommended for large datasets)
90
+ dataset = load_dataset("ShandaAI/Hive", streaming=True)
91
+ ```
92
+
93
+ ## ๐Ÿ“„ Dataset Description
94
+
95
+ **Hive** is a high-quality synthetic dataset designed for **Universal Sound Separation (USS)**. Unlike traditional methods relying on weakly-labeled in-the-wild data, Hive leverages an automated data collection pipeline to mine high-purity single-event segments from complex acoustic environments and synthesizes mixtures with semantically consistent constraints.
96
+
97
+ ### Key Features
98
+
99
+ - **Purity over Scale**: 2.4k hours achieving competitive performance with million-hour baselines (~0.2% data scale)
100
+ - **Single-label Clean Supervision**: Rigorous semantic-acoustic alignment eliminating co-occurrence noise
101
+ - **Semantically Consistent Mixing**: Logic-based co-occurrence matrix ensuring realistic acoustic scenes
102
+ - **High Fidelity**: 44.1kHz sample rate for high-quality audio
103
+
104
+ ### Dataset Scale
105
+
106
+ | Metric | Value |
107
+ |--------|-------|
108
+ | **Training Set Raw Audio** | 2,442 hours |
109
+ | **Val & Test Set Raw Audio** | 292 hours |
110
+ | **Mixed Samples** | 19.6M mixtures |
111
+ | **Total Mixed Duration** | ~22.4k hours |
112
+ | **Label Categories** | 283 classes |
113
+ | **Sample Rate** | 44.1 kHz |
114
+ | **Training Sample Duration** | 4 seconds |
115
+ | **Test Sample Duration** | 10 seconds |
116
+
117
+ ### Dataset Splits
118
+
119
+ | Split | Samples | Description |
120
+ |-------|---------|-------------|
121
+ | Train | 17.5M | Training mixtures (4s duration) |
122
+ | Validation | 1.75M | Validation mixtures |
123
+ | Test | 350k | Test mixtures (10s duration) |
124
+
125
+ ---
126
+
127
+ ## ๐Ÿ“‚ Dataset Structure
128
+
129
+ ### Directory Organization
130
+
131
+ ```
132
+ hive-datasets-parquet/
133
+ โ”œโ”€โ”€ README.md
134
+ โ”œโ”€โ”€ train/
135
+ โ”‚ โ””โ”€โ”€ data.parquet
136
+ โ”œโ”€โ”€ validation/
137
+ โ”‚ โ””โ”€โ”€ data.parquet
138
+ โ””โ”€โ”€ test/
139
+ โ””โ”€โ”€ data.parquet
140
+ ```
141
+
142
+ Each split contains a single Parquet file with all mixture metadata. The `num_sources` field indicates the number of sources (2-5) for each mixture.
143
+
144
+ ---
145
+
146
+ ## ๐Ÿ“‹ Data Fields
147
+
148
+ ### JSON Schema
149
+
150
+ Each JSON object contains complete generation parameters for reproducing a mixture sample:
151
+
152
+ ```python
153
+ {
154
+ "mix_id": "sample_00000003",
155
+ "split": "train",
156
+ "sample_rate": 44100,
157
+ "target_duration": 4.0,
158
+ "num_sources": 2,
159
+ "sources": {
160
+ "source_id": ["s1", "s2"],
161
+ "path": ["relative/path/to/audio1", "relative/path/to/audio2"],
162
+ "label": ["Ocean", "Rain"],
163
+ "crop_start_second": [1.396, 2.5],
164
+ "crop_end_second": [5.396, 6.5],
165
+ "chunk_start_second": [35.0, 20.0],
166
+ "chunk_end_second": [45.0, 30.0],
167
+ "rms_gain": [3.546, 2.1],
168
+ "snr_db": [0.0, -3.0],
169
+ "applied_weight": [3.546, 1.487]
170
+ },
171
+ "global_normalization_factor": 0.786,
172
+ "final_max_amplitude": 0.95
173
+ }
174
+ ```
175
+
176
+ ### Field Descriptions
177
+
178
+ #### 1. Basic Info Fields
179
+
180
+ | Field | Type | Description |
181
+ |-------|------|-------------|
182
+ | `mix_id` | string | Unique identifier for the mixture task |
183
+ | `split` | string | Dataset partition (`train` / `validation` / `test`) |
184
+ | `sample_rate` | int32 | Audio sample rate in Hz (44100) |
185
+ | `target_duration` | float64 | Target duration in seconds (4.0 for train, 10.0 for test) |
186
+ | `num_sources` | int32 | Number of audio sources in this mixture (2-5) |
187
+
188
+ #### 2. Source Information (`sources`)
189
+
190
+ Metadata required to reproduce the mixing process for each audio source. Stored in columnar format (dict of lists) for efficient Parquet storage:
191
+
192
+ | Field | Type | Description |
193
+ |-------|------|-------------|
194
+ | `source_id` | list[string] | Source identifiers (`s1`, `s2`, ...) |
195
+ | `path` | list[string] | Relative paths to the source audio files |
196
+ | `label` | list[string] | AudioSet ontology labels for each source |
197
+ | `chunk_start_second` | list[float64] | Start times (seconds) for reading from original audio files |
198
+ | `chunk_end_second` | list[float64] | End times (seconds) for reading from original audio files |
199
+ | `crop_start_second` | list[float64] | Precise start positions (seconds) for reproducible random extraction |
200
+ | `crop_end_second` | list[float64] | Precise end positions (seconds) for reproducible random extraction |
201
+ | `rms_gain` | list[float64] | Energy normalization coefficients: $\text{target\_rms} / \text{current\_rms}$ |
202
+ | `snr_db` | list[float64] | Signal-to-noise ratios in dB assigned to each source |
203
+ | `applied_weight` | list[float64] | Final scaling weights: $\text{rms\_gain} \times 10^{(\text{snr\_db} / 20)}$ |
204
+
205
+ #### 3. Mixing Parameters
206
+
207
+ Global processing parameters after combining multiple audio sources:
208
+
209
+ | Field | Type | Description |
210
+ |-------|------|-------------|
211
+ | `global_normalization_factor` | float64 | Anti-clipping scaling coefficient: $0.95 / \text{max\_val}$ |
212
+ | `final_max_amplitude` | float64 | Maximum amplitude threshold (0.95) to prevent bit-depth overflow |
213
+
214
+ ### Detailed Field Explanations
215
+
216
+ #### Cropping Logic
217
+ - `chunk_start/end_second`: Defines the reading interval from the original audio file
218
+ - `crop_start/end_second`: Records the precise random cropping position, ensuring exact reproducibility across runs
219
+
220
+ #### Energy Normalization (`rms_gain`)
221
+ Adjusts different audio sources to the same energy level:
222
+ $$\text{rms\_gain} = \frac{\text{target\_rms}}{\text{current\_rms}}$$
223
+
224
+ #### Signal-to-Noise Ratio (`snr_db`)
225
+ The SNR value assigned to each source, sampled from a predefined range using `random.uniform(snr_range[0], snr_range[1])`.
226
+
227
+ #### Applied Weight
228
+ The comprehensive scaling weight combining energy normalization and SNR adjustment:
229
+ $$\text{applied\_weight} = \text{rms\_gain} \times 10^{(\text{snr\_db} / 20)}$$
230
+
231
+ This is the final coefficient applied to the original waveform.
232
+
233
+ #### Global Normalization Factor
234
+ Prevents audio clipping after mixing:
235
+ $$\text{global\_normalization\_factor} = \frac{0.95}{\text{max\_val}}$$
236
+
237
+ Where `max_val` is the **peak amplitude (absolute value)** of the mixed signal.
238
+
239
+ ---
240
+
241
+ ## ๐Ÿ”ง Usage
242
+
243
+ ### Download Metadata
244
+
245
+ ```python
246
+ from datasets import load_dataset
247
+
248
+ # Load specific split and mixture type
249
+ dataset = load_dataset("ShandaAI/Hive", split="train")
250
+ ```
251
+
252
+ ### Generate Mixed Audio
253
+
254
+ Please refer to the [official GitHub repository](https://github.com/ShandaAI/Hive) for the complete audio generation pipeline.
255
+
256
+ ```bash
257
+ # Clone the repository
258
+ git clone https://github.com/ShandaAI/Hive.git
259
+ cd Hive/hive_dataset
260
+
261
+ # Generate mixtures from metadata
262
+ python mix_from_metadata/mix_from_metadata.py \
263
+ --metadata_dir /path/to/downloaded/metadata \
264
+ --output_dir ./hive_dataset \
265
+ --dataset_paths dataset_paths.json \
266
+ --num_processes 16
267
+ ```
268
+
269
+ ---
270
+
271
+ ## ๐Ÿ“š Source Datasets
272
+
273
+ Hive integrates **12 public datasets** to construct a long-tailed acoustic space:
274
+
275
+ | # | Dataset | Clips | Duration (h) | License |
276
+ |---|---------|-------|--------------|---------|
277
+ | 1 | BBC Sound Effects | 369,603 | 1,020.62 | Remix License |
278
+ | 2 | AudioSet | 326,890 | 896.61 | CC BY |
279
+ | 3 | VGGSound | 115,191 | 319.10 | CC BY 4.0 |
280
+ | 4 | MUSIC21 | 32,701 | 90.28 | YouTube Standard |
281
+ | 5 | FreeSound | 17,451 | 46.90 | CC0/BY/BY-NC |
282
+ | 6 | ClothoV2 | 14,759 | 38.19 | Non-Commercial Research |
283
+ | 7 | Voicebank-DEMAND | 12,376 | 9.94 | CC BY 4.0 |
284
+ | 8 | AVE | 3,054 | 6.91 | CC BY-NC-SA |
285
+ | 9 | SoundBible | 2,501 | 5.78 | CC BY 4.0 |
286
+ | 10 | DCASE | 1,969 | 5.46 | Academic Use |
287
+ | 11 | ESC50 | 1,433 | 1.99 | CC BY-NC 3.0 |
288
+ | 12 | FSD50K | 636 | 0.80 | Creative Commons |
289
+ | | **Total** | **898,564** | **2,442.60** | |
290
+
291
+ **Important Note**: This repository releases only **metadata** (JSON files containing mixing parameters and source references) for reproducibility. Users must independently download and prepare the source datasets according to their respective licenses.
292
+
293
+ ---
294
+
295
+ ## ๐Ÿ“– Citation
296
+
297
+ If you use this dataset, please cite:
298
+
299
+ ```bibtex
300
+ ```
301
+
302
+ ---
303
+
304
+ ## โš–๏ธ License
305
+
306
+ This dataset metadata is released under the **Apache License 2.0**.
307
+
308
+ Please note that the source audio files are subject to their original licenses. Users must comply with the respective licenses when using the source datasets.
309
+
310
+ ---
311
+
312
+ ## ๐Ÿ™ Acknowledgments
313
+
314
+ We extend our gratitude to the researchers and organizations who curated the foundational datasets that made Hive possible:
315
+
316
+ - **BBC Sound Effects** - Professional-grade recordings with broadcast-level fidelity
317
+ - **AudioSet** (Google) - Large-scale audio benchmark
318
+ - **VGGSound** (University of Oxford) - Real-world acoustic diversity
319
+ - **FreeSound** (MTG-UPF) - Rich crowdsourced soundscapes
320
+ - And all other contributing datasets
321
+
322
+ ---
323
+
324
+ ## ๐Ÿ“ฌ Contact
325
+
326
+ For questions or issues, please open an issue on the [GitHub repository](https://github.com/ShandaAI/Hive) or contact the authors.
asserts/logo.png ADDED

Git LFS Details

  • SHA256: cc439987217d8b8dbc3d9e48eab565f2e2e15199f68a4d245e5097d202093298
  • Pointer size: 132 Bytes
  • Size of remote file: 4.14 MB
test/data.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e9916476565357dcf675eee888232df3d574e82851ce93676b70050ccd477e9
3
+ size 15678703
train/data.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:074fab7cf25d100e0b4617706347816b5566b14d26ca94e14dece8522814b304
3
+ size 1106516997
validation/data.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ffc6c303b57cdaa2d2ec6d5edf317a2652d2200a3472fc7a31d2e795c89ceff
3
+ size 110792699