YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

NB: Performance can drop significantly when using the safetensors variant of this model. Load the model with use_safetensors=False for optimal performance

This is an encoder Language Model pre-trained from scratch on transcriptions of the archives of the Dutch East India Company. It is therefore a model specialized on Early Modern Dutch as used in the archive (1602โ€“1800). The model follows a RoBERTa architecture. It can be fine-tuned on any NLP task.

This version of the model is the best performing GloBERTise model when tested on binary event detection of the four I have pre-trained (in august 2025)

Comparison to other models: Adapted settings for 'num_training_steps' and 'num_warmup_steps' compared to GloBERTise-v01 and GloBERTise-v01-rerun, otherwise the same. Different seed compared to GloBERTise-rerun, same parameter settings.

See my GitHub repos

And a small presentation: https://docs.google.com/presentation/d/1gkg5hChWAMXA6mxfgFkkvIieWdj_17yKitwBkBNcJBo/edit?usp=sharing

Most important parameter settings:

learning rate 0.0003
betas [ 0.9, 0.98]
weight_decay 0.01
num_train_epochs 2
per_device_train_batch_size 40
gradient_accumulation_steps 10
fp16 true
Downloads last month
148
Safetensors
Model size
0.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for globalise/GloBERTise

Finetunes
1 model