GetmanY1 commited on
Commit
f124fd6
verified
1 Parent(s): 17ab9be

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -1
README.md CHANGED
@@ -20,7 +20,7 @@ The base model pre-trained on 16kHz sampled speech audio. When using the model m
20
 
21
  The Finnish Wav2Vec2 Base has the same architecture and uses the same training objective as the English and multilingual one described in [Paper](https://arxiv.org/abs/2006.11477). It is pre-trained on 158k hours of unlabeled Finnish speech, including [KAVI radio and television archive materials](https://kavi.fi/en/radio-ja-televisioarkistointia-vuodesta-2008/), Lahjoita puhetta (Donate Speech), Finnish Parliament, Finnish VoxPopuli.
22
 
23
- You can read more about the pre-trained model from [this paper](TODO). The training scripts are available on [GitHub](https://github.com/aalto-speech/large-scale-monolingual-speech-foundation-models).
24
 
25
  ## Intended uses & limitations
26
 
@@ -107,6 +107,22 @@ The pre-trained model was initialized with the following hyperparameters:
107
  - Pytorch 1.13.1+rocm5.2
108
  - Fairseq 0.12.2
109
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
110
  ## Team Members
111
 
112
  - Yaroslav Getman, [Hugging Face profile](https://huggingface.co/GetmanY1), [LinkedIn profile](https://www.linkedin.com/in/yaroslav-getman/)
 
20
 
21
  The Finnish Wav2Vec2 Base has the same architecture and uses the same training objective as the English and multilingual one described in [Paper](https://arxiv.org/abs/2006.11477). It is pre-trained on 158k hours of unlabeled Finnish speech, including [KAVI radio and television archive materials](https://kavi.fi/en/radio-ja-televisioarkistointia-vuodesta-2008/), Lahjoita puhetta (Donate Speech), Finnish Parliament, Finnish VoxPopuli.
22
 
23
+ You can read more about the pre-trained model from [this paper](https://www.isca-archive.org/interspeech_2025/getman25_interspeech.html). The training scripts are available on [GitHub](https://github.com/aalto-speech/large-scale-monolingual-speech-foundation-models).
24
 
25
  ## Intended uses & limitations
26
 
 
107
  - Pytorch 1.13.1+rocm5.2
108
  - Fairseq 0.12.2
109
 
110
+ ## Citation
111
+
112
+ If you use our models or scripts, please cite our article as:
113
+
114
+ ```bibtex
115
+ @inproceedings{getman25_interspeech,
116
+ title = {{Is your model big enough? Training and interpreting large-scale monolingual speech foundation models}},
117
+ author = {{Yaroslav Getman and Tam谩s Gr贸sz and Tommi Lehtonen and Mikko Kurimo}},
118
+ year = {{2025}},
119
+ booktitle = {{Interspeech 2025}},
120
+ pages = {{231--235}},
121
+ doi = {{10.21437/Interspeech.2025-46}},
122
+ issn = {{2958-1796}},
123
+ }
124
+ ```
125
+
126
  ## Team Members
127
 
128
  - Yaroslav Getman, [Hugging Face profile](https://huggingface.co/GetmanY1), [LinkedIn profile](https://www.linkedin.com/in/yaroslav-getman/)