File size: 1,647 Bytes
b0e2a85
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8adadf1
4c18d19
 
f91d994
 
ef45f2a
f91d994
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
---
dataset_info:
  features:
  - name: audio
    dtype: audio
  - name: label
    dtype:
      class_label:
        names:
          '0': I+have+one+now
          '1': I+only+have+one
  splits:
  - name: train
    num_bytes: 10168367.5
    num_examples: 535
  - name: test
    num_bytes: 1499291.5
    num_examples: 95
  - name: validation
    num_bytes: 1720511.5
    num_examples: 97
  download_size: 13330229
  dataset_size: 13388170.5
---
# Dataset Card for "have_one"
The dataset consists of utterances of *have one* that are cut either from an utterance of *I have one now*, or from an utterance
of *I only have one*.  The first tends to have prominence on *have*, while the second tends to have prominence on *one*. See
`github.com/MatsRooth/fiyou` on the methodology for finding the utterances on Youtube, and aligning and cutting them using
Kaldi.

To put such a dataset on huggingface hub, start with this directory structure, where the bottom directories contain wav files.  
```
have_one
└── data
    ├── I+have+one+now
    └── I+only+have+one
```
Run `have_one_hub.py` to create the dataset, using the generic Huggingface methodology for audio datasets. 

The dataset is used in the wav2vec2 binary classification model `MatsRooth/wav2vec2-base_have_one`.

Often cutting with a Kaldi phone alignment gives a snippet that includes part of preceding vowel, or has formant structure
in the start of /h/ that gives information about the preceding vowel. These vowels are different for
the two classes, and so classification can be based on this, as well as the intended prosodic difference.
This needs to be corrected.