rjzevallos commited on
Commit
d686ad4
·
verified ·
1 Parent(s): a1c6d62

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -30
README.md CHANGED
@@ -1,28 +1,36 @@
1
  ---
 
2
  language:
3
  - ca
 
 
 
 
4
  ---
5
 
6
- # PL-BERT-wordpiece-cat-multiaccent
 
 
7
 
8
  ## Overview
9
 
10
  <details>
11
  <summary>Click to expand</summary>
12
 
13
- - **Model type:** Phoneme-level Language Model (PL-BERT)
14
- - **Architecture:** AlBERT-base (12 layers, 768 hidden units, 12 attention heads)
15
- - **Language:** Catalan (multiple accents)
16
- - **License:** Apache 2.0
17
- - **Data:** Crowdsourced phonemized Catalan speech text
 
18
 
19
  </details>
20
 
21
  ---
22
 
23
- ## Model description
24
 
25
- **PL-BERT-wordpiece-cat-multiaccent** is a phoneme-level masked language model trained on Catalan text with diverse regional accents. It is based on the [PL-BERT architecture](https://github.com/yl4579/PL-BERT), which learns phoneme representations via a BERT-style masked language modeling objective.
26
 
27
  This model is designed to support **phoneme-based text-to-speech (TTS) systems**, including but not limited to [StyleTTS2](https://github.com/yl4579/StyleTTS2). Thanks to its Catalan-specific phoneme vocabulary and contextual embedding capabilities, it can serve as a phoneme encoder for any TTS architecture requiring phoneme-level features.
28
 
@@ -34,7 +42,7 @@ Features of our PL-BERT:
34
 
35
  ---
36
 
37
- ## Intended uses and limitations
38
 
39
  ### Intended uses
40
 
@@ -50,7 +58,7 @@ Features of our PL-BERT:
50
 
51
  ---
52
 
53
- ## How to use (with StyleTTS2)
54
 
55
  Here is an example of how to use this model within the StyleTTS2 framework:
56
 
@@ -76,7 +84,7 @@ Note: Although this example uses StyleTTS2, the model is compatible with other T
76
 
77
  ---
78
 
79
- ## Training
80
 
81
  ### Training data
82
 
@@ -110,49 +118,46 @@ Other parameters:
110
  - Token mask: M
111
  - Word separator ID: 102
112
 
113
- ---
114
-
115
 
116
- ## Evaluation
117
 
118
  The model has not been benchmarked via perplexity or extrinsic evaluation, but has been successfully integrated into TTS pipelines such as StyleTTS2, where it enables the synthesis of Catalan with regional accent variation.
119
 
120
  ---
121
 
 
122
  ## Citation
123
 
124
  If this code contributes to your research, please cite the work:
125
 
126
  ```
127
- @misc{LangtechVeu2025plbertwordpiececatmultiaccent,
128
- title={PL-BERT-wordpiece-cat-multiaccent},
129
  author={Rodolfo Zevallos, Jose Giraldo and Carme Armentano-Oller},
130
  organization={Barcelona Supercomputing Center},
131
- url={https://huggingface.co/langtech-veu/PL-BERT-wordpiece-cat-multiaccent},
132
  year={2025}
133
  }
134
  ```
135
 
136
- ## Additional information
137
 
138
- ### Contact
139
 
140
- For questions or feedback, please contact:
141
- rodolfo.zevallos@bsc.es
142
 
143
- ### License
144
 
145
- Distributed under the Apache License, Version 2.0: https://www.apache.org/licenses/LICENSE-2.0
 
146
 
147
- ### Funding
148
- This work is funded by the Ministerio para la Transformación Digital y de la Función Pública - Funded by EU – NextGenerationEU within the framework of the project Desarrollo de Modelos ALIA.
149
 
 
150
 
151
- ### Disclaimer
152
 
153
- <details>
154
- <summary>Click to expand</summary>
155
 
156
- This model is released for research and educational use. It may exhibit biases or limitations based on training data characteristics. Users are responsible for ensuring appropriate use in deployed systems and for complying with all applicable regulations.
 
157
 
158
- </details>
 
1
  ---
2
+ license: apache-2.0
3
  language:
4
  - ca
5
+ tags:
6
+ - TTS
7
+ - PL-BERT
8
+ - barcelona-supercomputing-center
9
  ---
10
 
11
+
12
+ # PL-BERT-wp-ca
13
+
14
 
15
  ## Overview
16
 
17
  <details>
18
  <summary>Click to expand</summary>
19
 
20
+ - [Model Description](#model-description)
21
+ - [Intended Uses and Limitations](#intended-uses-and-limitations)
22
+ - [How to Get Started with the Model](#how-to-get-started-with-the-model)
23
+ - [Training Details](#training-details)
24
+ - [Citation](#citation)
25
+ - [Additional information](#additional-information)
26
 
27
  </details>
28
 
29
  ---
30
 
31
+ ## Model Description
32
 
33
+ **PL-BERT-wp-ca** is a phoneme-level masked language model trained on Catalan text with diverse regional accents. It is based on the [PL-BERT architecture](https://github.com/yl4579/PL-BERT), which learns phoneme representations via a BERT-style masked language modeling objective.
34
 
35
  This model is designed to support **phoneme-based text-to-speech (TTS) systems**, including but not limited to [StyleTTS2](https://github.com/yl4579/StyleTTS2). Thanks to its Catalan-specific phoneme vocabulary and contextual embedding capabilities, it can serve as a phoneme encoder for any TTS architecture requiring phoneme-level features.
36
 
 
42
 
43
  ---
44
 
45
+ ## Intended Uses and Limitations
46
 
47
  ### Intended uses
48
 
 
58
 
59
  ---
60
 
61
+ ## How to Get Started with the Model
62
 
63
  Here is an example of how to use this model within the StyleTTS2 framework:
64
 
 
84
 
85
  ---
86
 
87
+ ## Training Details
88
 
89
  ### Training data
90
 
 
118
  - Token mask: M
119
  - Word separator ID: 102
120
 
 
 
121
 
122
+ ### Evaluation
123
 
124
  The model has not been benchmarked via perplexity or extrinsic evaluation, but has been successfully integrated into TTS pipelines such as StyleTTS2, where it enables the synthesis of Catalan with regional accent variation.
125
 
126
  ---
127
 
128
+
129
  ## Citation
130
 
131
  If this code contributes to your research, please cite the work:
132
 
133
  ```
134
+ @misc{zevallosbertwpca,
135
+ title={PL-BERT-wp-ca},
136
  author={Rodolfo Zevallos, Jose Giraldo and Carme Armentano-Oller},
137
  organization={Barcelona Supercomputing Center},
138
+ url={https://huggingface.co/langtech-veu/PL-BERT-wp-ca},
139
  year={2025}
140
  }
141
  ```
142
 
143
+ ## Additional Information
144
 
 
145
 
146
+ ### Author
 
147
 
148
+ The [Language Technologies Laboratory](https://huggingface.co/BSC-LT) of the [Barcelona Supercomputing Center](https://www.bsc.es/) by [Rodolfo Zevallos](https://huggingface.co/rjzevallos).
149
 
150
+ ### Contact
151
+ For further information, please send an email to <langtech@bsc.es>.
152
 
153
+ ### Copyright
154
+ Copyright(c) 2025 by Language Technologies Laboratory, Barcelona Supercomputing Center.
155
 
156
+ ### License
157
 
158
+ [Apache-2.0](https://www.apache.org/licenses/LICENSE-2.0)
159
 
 
 
160
 
161
+ ### Funding
162
+ This work is funded by the Ministerio para la Transformación Digital y de la Función Pública - Funded by EU – NextGenerationEU within the framework of the project Desarrollo de Modelos ALIA.
163