Upload 3 files
Browse filesPython code that creates the dataset.
- down_on_copy.py +67 -0
- down_on_create.py +80 -0
- down_on_hub.py +14 -0
down_on_copy.py
ADDED
|
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import datasets
|
| 2 |
+
import os
|
| 3 |
+
import shutil
|
| 4 |
+
#from datasets import load_dataset
|
| 5 |
+
train0, validation0, test0 = datasets.load_dataset("superb", "ks", split=["train","validation","test"])
|
| 6 |
+
|
| 7 |
+
labels = train0.features["label"].names
|
| 8 |
+
label2id = {x: labels.index(x) for x in labels}
|
| 9 |
+
# id2label = {str(id): label for label, id in label2id.items()}
|
| 10 |
+
|
| 11 |
+
down_id = label2id['down']
|
| 12 |
+
on_id = label2id['on']
|
| 13 |
+
|
| 14 |
+
train0_down = train0.filter(lambda example: example['label'] == down_id)
|
| 15 |
+
train0_on = train0.filter(lambda example: example['label'] == on_id)
|
| 16 |
+
|
| 17 |
+
shutil.rmtree('/home/mr249/ac_h/down_on/data/train/down',ignore_errors=True)
|
| 18 |
+
os.mkdir('/home/mr249/ac_h/down_on/data/train/down')
|
| 19 |
+
shutil.rmtree('/home/mr249/ac_h/down_on/data/train/on',ignore_errors=True)
|
| 20 |
+
os.mkdir('/home/mr249/ac_h/down_on/data/train/on')
|
| 21 |
+
|
| 22 |
+
for e in train0_down:
|
| 23 |
+
p = e['audio']['path']
|
| 24 |
+
newpath = '/home/mr249/ac_h/down_on/data/train/down/{}'.format(p.split('/')[len(p.split('/')) - 1])
|
| 25 |
+
os.link(p,newpath)
|
| 26 |
+
|
| 27 |
+
for e in train0_on:
|
| 28 |
+
p = e['audio']['path']
|
| 29 |
+
newpath = '/home/mr249/ac_h/down_on/data/train/on/{}'.format(p.split('/')[len(p.split('/')) - 1])
|
| 30 |
+
os.link(p,newpath)
|
| 31 |
+
|
| 32 |
+
validation0_down = validation0.filter(lambda example: example['label'] == down_id)
|
| 33 |
+
validation0_on = validation0.filter(lambda example: example['label'] == on_id)
|
| 34 |
+
|
| 35 |
+
shutil.rmtree('/home/mr249/ac_h/down_on/data/validation/down',ignore_errors=True)
|
| 36 |
+
os.mkdir('/home/mr249/ac_h/down_on/data/validation/down')
|
| 37 |
+
shutil.rmtree('/home/mr249/ac_h/down_on/data/validation/on',ignore_errors=True)
|
| 38 |
+
os.mkdir('/home/mr249/ac_h/down_on/data/validation/on')
|
| 39 |
+
|
| 40 |
+
for e in validation0_down:
|
| 41 |
+
p = e['audio']['path']
|
| 42 |
+
newpath = '/home/mr249/ac_h/down_on/data/validation/down/{}'.format(p.split('/')[len(p.split('/')) - 1])
|
| 43 |
+
os.link(p,newpath)
|
| 44 |
+
|
| 45 |
+
for e in validation0_on:
|
| 46 |
+
p = e['audio']['path']
|
| 47 |
+
newpath = '/home/mr249/ac_h/down_on/data/validation/on/{}'.format(p.split('/')[len(p.split('/')) - 1])
|
| 48 |
+
os.link(p,newpath)
|
| 49 |
+
|
| 50 |
+
test0_down = test0.filter(lambda example: example['label'] == down_id)
|
| 51 |
+
test0_on = test0.filter(lambda example: example['label'] == on_id)
|
| 52 |
+
|
| 53 |
+
shutil.rmtree('/home/mr249/ac_h/down_on/data/test/down',ignore_errors=True)
|
| 54 |
+
os.mkdir('/home/mr249/ac_h/down_on/data/test/down')
|
| 55 |
+
shutil.rmtree('/home/mr249/ac_h/down_on/data/test/on',ignore_errors=True)
|
| 56 |
+
os.mkdir('/home/mr249/ac_h/down_on/data/test/on')
|
| 57 |
+
|
| 58 |
+
for e in test0_down:
|
| 59 |
+
p = e['audio']['path']
|
| 60 |
+
newpath = '/home/mr249/ac_h/down_on/data/test/down/{}'.format(p.split('/')[len(p.split('/')) - 1])
|
| 61 |
+
os.link(p,newpath)
|
| 62 |
+
|
| 63 |
+
for e in test0_on:
|
| 64 |
+
p = e['audio']['path']
|
| 65 |
+
newpath = '/home/mr249/ac_h/down_on/data/test/on/{}'.format(p.split('/')[len(p.split('/')) - 1])
|
| 66 |
+
os.link(p,newpath)
|
| 67 |
+
|
down_on_create.py
ADDED
|
@@ -0,0 +1,80 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import datasets
|
| 2 |
+
#from datasets import load_dataset
|
| 3 |
+
train0, validation0, test0 = datasets.load_dataset("superb", "ks", split=["train","validation","test"])
|
| 4 |
+
|
| 5 |
+
labels = train0.features["label"].names
|
| 6 |
+
label2id = {x: labels.index(x) for x in labels}
|
| 7 |
+
id2label = {str(id): label for label, id in label2id.items()}
|
| 8 |
+
|
| 9 |
+
down_id = label2id['down']
|
| 10 |
+
on_id = label2id['on']
|
| 11 |
+
|
| 12 |
+
# This filters 51094 rows to 3706 rows, with features ['file', 'audio', 'label']
|
| 13 |
+
train1 = train0.filter(lambda example: example['label'] == down_id or example['label'] == on_id)
|
| 14 |
+
|
| 15 |
+
# 521 rows
|
| 16 |
+
validation1 = validation0.filter(lambda example: example['label'] == down_id or example['label'] == on_id)
|
| 17 |
+
# 499 rows
|
| 18 |
+
test1 = test0.filter(lambda example: example['label'] == down_id or example['label'] == on_id)
|
| 19 |
+
|
| 20 |
+
train1.to_csv('/home/mr249/ac_h/do1/tmp/train1.csv')
|
| 21 |
+
validation1.to_csv('/home/mr249/ac_h/do1/tmp/validation1.csv')
|
| 22 |
+
test1.to_csv('/home/mr249/ac_h/do1/tmp/test1.csv')
|
| 23 |
+
|
| 24 |
+
# Fix the labels. TODO Put this in this python program.
|
| 25 |
+
# This is kindof in the wrong location?
|
| 26 |
+
# /home/mr249/ac_h/do1
|
| 27 |
+
# cat tmp/train1.csv | awk -f ../script/fix-down-on-labels.awk > train.csv
|
| 28 |
+
# cat tmp/validation1.csv | awk -f ../script/fix-down-on-labels.awk > validation.csv
|
| 29 |
+
# cat tmp/test1.csv | awk -f ../script/fix-down-on-labels.awk > test.csv
|
| 30 |
+
|
| 31 |
+
# Create new datasets from the csv files
|
| 32 |
+
|
| 33 |
+
train2 = datasets.Dataset.from_csv('/home/mr249/ac_h/do1/train.csv','train')
|
| 34 |
+
validation2 = datasets.Dataset.from_csv('/home/mr249/ac_h/do1/validation.csv','validation')
|
| 35 |
+
test2 = datasets.Dataset.from_csv('/home/mr249/ac_h/do1/test.csv','test')
|
| 36 |
+
|
| 37 |
+
# Above, the second argument names the split. It may not be relevant.
|
| 38 |
+
# validation2.split
|
| 39 |
+
# NamedSplit('validation')
|
| 40 |
+
|
| 41 |
+
# Add the label names. It does not change the examples, which have ints as labels.
|
| 42 |
+
new_features = train2.features.copy()
|
| 43 |
+
new_features["label"] = datasets.ClassLabel(names=['down', 'on'],id=None)
|
| 44 |
+
train2 = train2.cast(new_features)
|
| 45 |
+
validation2 = validation2.cast(new_features)
|
| 46 |
+
test2 = test2.cast(new_features)
|
| 47 |
+
|
| 48 |
+
# Combine them into a DatasetDict.
|
| 49 |
+
down_on = datasets.DatasetDict({
|
| 50 |
+
"train": train2,
|
| 51 |
+
"validation": validation2,
|
| 52 |
+
"test": test2,
|
| 53 |
+
})
|
| 54 |
+
|
| 55 |
+
# Save to disk
|
| 56 |
+
down_on.save_to_disk('/home/mr249/ac_h/down_on')
|
| 57 |
+
|
| 58 |
+
# Load from disk
|
| 59 |
+
# However, we prefer to load from the hub.
|
| 60 |
+
# down_on2 = datasets.load_from_disk('/home/mr249/ac_h/down_on')
|
| 61 |
+
|
| 62 |
+
# Push to hub
|
| 63 |
+
# See https://huggingface.co/docs/datasets/upload_dataset
|
| 64 |
+
from huggingface_hub import login
|
| 65 |
+
login()
|
| 66 |
+
# It prompts for the token
|
| 67 |
+
down_on.push_to_hub("MatsRooth/down_on",private=False,embed_external_files=True)
|
| 68 |
+
|
| 69 |
+
# Now the dataset can be loaded from the hub!
|
| 70 |
+
train3 = load_dataset("MatsRooth/down_on", split="train")
|
| 71 |
+
|
| 72 |
+
# When running training, the names are wrong
|
| 73 |
+
# FileNotFoundError: [Errno 2] No such file or directory: "{'bytes': None, 'path':
|
| 74 |
+
# '/home/mr249/.cache/huggingface/datasets/downloads/extracted/5836831ec57281eff9b
|
| 75 |
+
# 1882385bf370d016058e6cba1fd7ff1dcb68cd8cddefd/down/28ed6bc9_nohash_4.wav'}"
|
| 76 |
+
|
| 77 |
+
|
| 78 |
+
# train3.features['label']
|
| 79 |
+
# ClassLabel(names=['down', 'on'], id=None)
|
| 80 |
+
|
down_on_hub.py
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# See https://huggingface.co/docs/datasets/audio_dataset
|
| 2 |
+
from datasets import load_dataset
|
| 3 |
+
|
| 4 |
+
# Upload the down_on dataset to huggingfacehub
|
| 5 |
+
|
| 6 |
+
# This magically does the right thing to create the dataset, including
|
| 7 |
+
# setting the label names and IDs.
|
| 8 |
+
ds = load_dataset("audiofolder", data_dir="down_on/data")
|
| 9 |
+
|
| 10 |
+
from huggingface_hub import login
|
| 11 |
+
login()
|
| 12 |
+
ds.push_to_hub("MatsRooth/down_on",private=False,embed_external_files=True)
|
| 13 |
+
|
| 14 |
+
|