have_one / README.md
MatsRooth's picture
Update README.md
29a105f verified
---
dataset_info:
features:
- name: audio
dtype: audio
- name: label
dtype:
class_label:
names:
'0': I+have+one+now
'1': I+only+have+one
splits:
- name: train
num_bytes: 10168367.5
num_examples: 535
- name: test
num_bytes: 1499291.5
num_examples: 95
- name: validation
num_bytes: 1720511.5
num_examples: 97
download_size: 13330229
dataset_size: 13388170.5
---
# Dataset Card for "have_one"
The dataset consists of utterances of *have one* that are cut either from an utterance of *I have one now*, or from an utterance
of *I only have one*. The first tends to have prominence on *have*, while the second tends to have prominence on *one*. See
`github.com/MatsRooth/fiyou` on the methodology for finding the utterances on Youtube, and aligning and cutting them using
Kaldi.
To put such a dataset on huggingface hub, start with this directory structure, where the bottom directories contain wav files.
```
have_one
└── data
├── I+have+one+now
└── I+only+have+one
```
Run `have_one_hub.py` to create the dataset, using the generic Huggingface methodology for audio datasets.
The dataset is used in the wav2vec2 binary classification model `MatsRooth/wav2vec2-base_have_one`.
Often cutting with a Kaldi phone alignment gives a snippet that includes part of preceding vowel, or has formant structure
in the start of /h/ that gives information about the preceding vowel. These vowels are different for
the two classes, and so classification can be based on this, as well as the intended prosodic difference.
This needs to be corrected.