{ "dataset_name": "test4", "description": "Merged speech dataset containing 345 segments from 2 source datasets", "features": { "audio": { "_type": "Audio", "sampling_rate": 16000 }, "text": { "_type": "Value", "dtype": "string" }, "speaker_id": { "_type": "Value", "dtype": "string" }, "emotion": { "_type": "Value", "dtype": "string" }, "language": { "_type": "Value", "dtype": "string" } }, "splits": { "train": { "name": "train", "num_examples": 345 } }, "total_segments": 345, "speakers": [ "speaker_1", "speaker_2", "speaker_6", "speaker_5", "speaker_3", "speaker_0", "speaker_4" ], "emotions": [ "neutral", "sad", "angry", "happy" ], "languages": [ "en" ], "original_datasets": [ "14a1fc0e-763a-4576-b2fb-9844c39946f2", "962aa070-c0d3-4064-944e-d7131ca38414" ] }