Icelandic TEAMS-Small
This model was pre-trained on the 2022 version of the Icelandic Gigaword Corpus using the TensorFlow Model Garden. Its pre-training configuration is identical to that of the original TEAMS-Small model, except for the use of a Unigram tokenizer with a vocabulary size of 64,105.
TEAMS is a variant of the ELECTRA architecture, described in Training ELECTRA Augmented with Multi-word Selection. While architecturally equivalent to ELECTRA, TEAMS introduces an additional pre-training objective.
Acknowledgments
This research was supported with Cloud TPUs from Google's TPU Research Cloud (TRC).
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support