abnerh commited on
Commit
33a849a
·
verified ·
1 Parent(s): f5253f8

Create README

Browse files
Files changed (1) hide show
  1. README.md +49 -0
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ tags:
4
+ - audio
5
+ - speech
6
+ - phonology
7
+ - wav2vec2
8
+ - multilingual
9
+ - pytorch-lightning
10
+ language:
11
+ - en
12
+ - es
13
+ - de
14
+ - cs
15
+ pipeline_tag: audio-classification
16
+ ---
17
+
18
+ # PhonoQ 2.0 – Multilingual
19
+
20
+ This repository hosts the **multilingual checkpoint** for **PhonoQ 2.0**, a modernized successor to the original PhonoQ system:
21
+ https://github.com/TAriasVergara/PhonoQ
22
+
23
+ PhonoQ 2.0 outputs **framewise probability distributions** over phonological heads from raw speech audio, built on a self-supervised speech encoder (e.g., wav2vec 2.0 / HuBERT).
24
+
25
+ ## What this model outputs
26
+
27
+ Given an input audio file, the model produces **framewise head probabilities** for:
28
+
29
+ - **Manner** (9 classes)
30
+ - **Vowel height** (3 classes)
31
+ - **Vowel backness** (3 classes)
32
+ - **Place of articulation** (5 classes)
33
+ - **Voicing** (2 classes)
34
+
35
+ Outputs are aligned to the encoder frame rate and returned as probabilities (not hard labels).
36
+
37
+ ## How to use
38
+
39
+ This checkpoint is intended to be used with the PhonoQ 2.0 inference code:
40
+ https://github.com/abnerLing/PhonoQ-2.0
41
+
42
+ ### 1) Install PhonoQ 2.0 (from GitHub)
43
+
44
+ Follow the installation instructions in the GitHub repository (PyTorch is required).
45
+
46
+ ### 2) Download this checkpoint
47
+
48
+ ```bash
49
+ wget https://huggingface.co/abnerh/phonoq-2.0-multilingual/resolve/main/best.ckpt