| --- |
| license: openrail |
| language: |
| - nl |
| tags: |
| - wav2vec2 |
| - self-supervised |
| - pretraining |
| - speech |
| - audio |
| --- |
| # Wav2Vec2-NL |
| A Dutch Wav2Vec2-base model, pre-trained on 831 hours of exclusively Dutch speech. |
|
|
| Pre-training data was extracted from a combination of: |
| - the [Spoken Dutch Corpus](https://taalmaterialen.ivdnt.org/wp-content/uploads/documentatie/cgn_website/doc_English/topics/index.htm) (537 hours; incl. spontaneous conversations, interviews, read speech and news reports) |
| - the Dutch component of [Multilingual LibriSpeech](https://www.openslr.org/94/) (211 hours; audiobook segments) |
| - the Dutch subset of the [CommonVoice 16.1](https://huggingface.co/datasets/mozilla-foundation/common_voice_16_1) corpus (83 hours; read aloud speech) |
|
|
| More information, incl. the training manifest and configuration is available in the [Wav2Vec2-NL repository on Zenodo](http://doi.org/10.5281/zenodo.15550628). |
|
|
| Analyses of Dutch phonetic and lexical features encoded in Wav2Vec2-NL hidden states are reported in the paper [What do self-supervised speech models know about Dutch? Analyzing advantages of language-specific pre-training](https://arxiv.org/abs/2506.00981) (Interspeech 2025; see full citation [below](#citation)). |
|
|
| Note: This model does not have a tokenizer as it was pretrained on audio alone. In order to use this model for speech recognition, a tokenizer should be created and the model should be fine-tuned on labeled text data. Check out [this blog](https://huggingface.co/blog/fine-tune-wav2vec2-english) for an explanation of fine-tuning Wav2Vec2 models on HuggingFace. |
|
|
| # Usage |
| ```python |
| from transformers import Wav2Vec2FeatureExtractor, Wav2Vec2Model |
| |
| feature_extractor = Wav2Vec2FeatureExtractor.from_pretrained('amsterdamNLP/Wav2Vec2-NL') |
| model = Wav2Vec2Model.from_pretrained('amsterdamNLP/Wav2Vec2-NL') |
| ``` |
|
|
| # Citation |
| The _Wav2Vec2-NL_ model was published as part of: |
| de Heer Kloots, M., Mohebbi, H., Pouw, C., Shen, G., Zuidema, W., Bentum, M. (2025). What do self-supervised speech models know about Dutch? Analyzing advantages of language-specific pre-training. _Proc. INTERSPEECH 2025_. https://doi.org/10.48550/arXiv.2506.00981 |
|
|
| BibTex entry: |
| ```bibtex |
| @inproceedings{deheerkloots25_interspeech, |
| title = {What do self-supervised speech models know about Dutch? Analyzing advantages of language-specific pre-training}, |
| author = {Marianne {de Heer Kloots} and Hosein Mohebbi and Charlotte Pouw and Gaofei Shen and Willem Zuidema and Martijn Bentum}, |
| year = {2025}, |
| booktitle = {Interspeech 2025}, |
| doi = {10.21437/Interspeech.2025-1526}, |
| } |
| ``` |