Stanwang1210 commited on
Commit
d484283
·
1 Parent(s): 7c33608

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +16 -11
README.md CHANGED
@@ -2,12 +2,14 @@
2
  configs:
3
  - config_name: default
4
  data_files:
5
- - split: train
6
- path: data/train-*
7
- - split: test
8
- path: data/test-*
9
- - split: validation
10
- path: data/validation-*
 
 
11
  dataset_info:
12
  features:
13
  - name: text
@@ -18,17 +20,20 @@ dataset_info:
18
  sequence:
19
  sequence: int64
20
  splits:
21
- - name: train
22
  num_bytes: 1166450829
23
  num_examples: 28539
24
- - name: test
25
  num_bytes: 62745230
26
  num_examples: 2620
27
- - name: validation
 
 
 
28
  num_bytes: 62578176
29
  num_examples: 2703
30
- download_size: 204598836
31
- dataset_size: 1291774235
32
  ---
33
  # Dataset Card for "speech_tokenizer_16k"
34
 
 
2
  configs:
3
  - config_name: default
4
  data_files:
5
+ - split: train.clean.100
6
+ path: data/train.clean.100-*
7
+ - split: test.clean
8
+ path: data/test.clean-*
9
+ - split: train.clean.360
10
+ path: data/train.clean.360-*
11
+ - split: validation.clean
12
+ path: data/validation.clean-*
13
  dataset_info:
14
  features:
15
  - name: text
 
20
  sequence:
21
  sequence: int64
22
  splits:
23
+ - name: train.clean.100
24
  num_bytes: 1166450829
25
  num_examples: 28539
26
+ - name: test.clean
27
  num_bytes: 62745230
28
  num_examples: 2620
29
+ - name: train.clean.360
30
+ num_bytes: 4216515060
31
+ num_examples: 104014
32
+ - name: validation.clean
33
  num_bytes: 62578176
34
  num_examples: 2703
35
+ download_size: 872101305
36
+ dataset_size: 5508289295
37
  ---
38
  # Dataset Card for "speech_tokenizer_16k"
39