Icelandic ConvBERT-Base

This model was pretrained on the Icelandic Gigaword Corpus, which contains approximately 1.69B tokens, using default settings. The model uses a WordPiece tokenizer with a vocabulary size of 32,105.

Acknowledgments

This research was supported with Cloud TPUs from Google's TPU Research Cloud (TRC).

This project was funded by the Language Technology Programme for Icelandic 2019-2023. The programme, which is managed and coordinated by Almannar贸mur, is funded by the Icelandic Ministry of Education, Science and Culture.

Downloads last month
13
Safetensors
Model size
0.1B params
Tensor type
I64
F32
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support