dataset_info:
features:
- name: audio
dtype: audio
- name: label
dtype:
class_label:
names:
'0': I+have+one+now
'1': I+only+have+one
splits:
- name: train
num_bytes: 10168367.5
num_examples: 535
- name: test
num_bytes: 1499291.5
num_examples: 95
- name: validation
num_bytes: 1720511.5
num_examples: 97
download_size: 13330229
dataset_size: 13388170.5
Dataset Card for "have_one"
The dataset consists of utterances of have one that are cut either from an utterance of I have one now, or from an utterance
of I only have one. The first tends to have prominence on have, while the second tends to have prominence on one. See
github.com/MatsRooth/fiyou on the methodology for finding the utterances on Youtube, and aligning and cutting them using
Kaldi.
To put such a dataset on huggingface hub, start with this directory structure, where the bottom directories contain wav files.
have_one
└── data
├── I+have+one+now
└── I+only+have+one
Run have_one_hub.py to create the dataset, using the generic Huggingface methodology for audio datasets.
The dataset is used in the wav2vec2 binary classification model MatsRooth/wav2vec2-base_have_one.
Often cutting with a Kaldi phone alignment gives a snippet that includes part of preceding vowel, or has formant structure in the start of /h/ that gives information about the preceding vowel. These vowels are different for the two classes, and so classification can be based on this, as well as the intended prosodic difference. This needs to be corrected.