ASLP-lab commited on
Commit
8d95ee2
·
verified ·
1 Parent(s): b35d84f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -34
README.md CHANGED
@@ -30,46 +30,47 @@ WenetSpeech-Chuan
30
 
31
  #### audio_labels/wav_utt_id.jsonl:
32
  {
33
- "wav_utt_id_timestamp": 以 转化为wav后的长音频id_时间戳信息 作为切分后的短音频id (type: str),
34
- "wav_utt_id_timestamp_path": 短音频数据路径 (type: str),
35
- "audio_clip_id": 该段短音频在长音频中的切分顺序编号,
36
- "timestamp": 时间戳信息,
37
- "wvmos_score": wvmos分数,衡量音频片段质量 (type: float),
38
- "text": 对应时间戳的音频片段的抄本 (type: str),
39
- "text_punc": 带标点的抄本 (type: str),
40
- "spk_num": 音频片段说话人个数,single/multi (type: str)
41
- "confidence": 抄本置信度 (type: float),
42
- "emotion": 说话人情感标签 (type: str,eg: 愤怒),
43
- "age": 说话人年龄标签 (type: int范围, eg: 中年(36~59)),
44
- "gender": 说话人性别标签 (type: str,eg: 男/女),
45
- }
46
 
47
  ### Data sample(EN):
48
 
49
  #### metadata.jsonl
50
- {
51
- "utt_id": Original long audio ID,
52
- "wav_utt_id": Converted long audio ID after transforming to WAV format,
53
- "source_audio_path": Path to the original long audio file,
54
- "audio_labels": Path to the label file of short audio segments cut from the converted long audio,
55
- "url": Download link for the original long audio
56
- }
57
 
58
 
59
  #### audio_labels/wav_utt_id.jsonl:
60
- {
61
- "wav_utt_id_timestamp": Short audio segment ID, composed of the converted long audio ID + timestamp information (type: str),
62
- "wav_utt_id_timestamp_path": Path to the short audio data (type: str),
63
- "audio_clip_id": Sequence number of this short segment within the long audio,
64
- "timestamp": Timestamp information,
65
- "wvmos_score": WVMOS score, measuring the quality of the audio segment (type: float),
66
- "text": Transcript of the audio segment corresponding to the timestamp (type: str),
67
- "text_punc": Transcript with punctuation (type: str),
68
- "spk_num": Number of speakers in the audio segment, single/multi (type: str),
69
- "confidence": Confidence score of the transcript (type: float),
70
- "emotion": Speaker’s emotion label (type: str, e.g., anger),
71
- "age": Speaker’s age label (type: int range, e.g., middle-aged (36–59)),
72
- "gender": Speaker’s gender label (type: str, e.g., male/female)
73
- }
 
74
 
75
 
 
30
 
31
  #### audio_labels/wav_utt_id.jsonl:
32
  {
33
+ "wav_utt_id_timestamp": 以 转化为wav后的长音频id_时间戳信息 作为切分后的短音频id (type: str), <br>
34
+ "wav_utt_id_timestamp_path": 短音频数据路径 (type: str), <br>
35
+ "audio_clip_id": 该段短音频在长音频中的切分顺序编号, <br>
36
+ "timestamp": 时间戳信息, <br>
37
+ "wvmos_score": wvmos分数,衡量音频片段质量 (type: float), <br>
38
+ "text": 对应时间戳的音频片段的抄本 (type: str), <br>
39
+ "text_punc": 带标点的抄本 (type: str), <br>
40
+ "spk_num": 音频片段说话人个数,single/multi (type: str) <br>
41
+ "confidence": 抄本置信度 (type: float), <br>
42
+ "emotion": 说话人情感标签 (type: str,eg: 愤怒), <br>
43
+ "age": 说话人年龄标签 (type: int范围, eg: 中年(36~59)), <br>
44
+ "gender": 说话人性别标签 (type: str,eg: 男/女), <br>
45
+ } <br>
46
 
47
  ### Data sample(EN):
48
 
49
  #### metadata.jsonl
50
+ { <br>
51
+ "utt_id": Original long audio ID, <br>
52
+ "wav_utt_id": Converted long audio ID after transforming to WAV format, <br>
53
+ "source_audio_path": Path to the original long audio file, <br>
54
+ "audio_labels": Path to the label file of short audio segments cut from the converted long audio, <br>
55
+ "url": Download link for the original long audio <br>
56
+ } <br>
57
 
58
 
59
  #### audio_labels/wav_utt_id.jsonl:
60
+
61
+ { <br>
62
+ "wav_utt_id_timestamp": Short audio segment ID, composed of the converted long audio ID + timestamp information (type: str), <br>
63
+ "wav_utt_id_timestamp_path": Path to the short audio data (type: str), <br>
64
+ "audio_clip_id": Sequence number of this short segment within the long audio, <br>
65
+ "timestamp": Timestamp information, <br>
66
+ "wvmos_score": WVMOS score, measuring the quality of the audio segment (type: float), <br>
67
+ "text": Transcript of the audio segment corresponding to the timestamp (type: str), <br>
68
+ "text_punc": Transcript with punctuation (type: str), <br>
69
+ "spk_num": Number of speakers in the audio segment, single/multi (type: str), <br>
70
+ "confidence": Confidence score of the transcript (type: float), <br>
71
+ "emotion": Speaker’s emotion label (type: str, e.g., anger), <br>
72
+ "age": Speaker’s age label (type: int range, e.g., middle-aged (36–59)), <br>
73
+ "gender": Speaker’s gender label (type: str, e.g., male/female) <br>
74
+ } <br>
75
 
76