How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("fill-mask", model="mascIT/bert-tiny-ita")
# Load model directly
from transformers import AutoTokenizer, AutoModelForMaskedLM

tokenizer = AutoTokenizer.from_pretrained("mascIT/bert-tiny-ita")
model = AutoModelForMaskedLM.from_pretrained("mascIT/bert-tiny-ita")
Quick Links

bert-tiny-ita is an italian foundational model (based on bert-tiny) pretrained from scratch on 20k italian Wikipedia articles and on a wide collection of italian words and dictionary definitions. It uses 512 context window size.

The project is still a work in progress, new versions will come with time.

Use it as a foundational model to be finetuned for specific italian tasks.

Training

  • epochs: 250
  • lr: 1e-5
  • optim: AdamW
  • weight_decay: 1e-4

Eval

  • perplexity: 45 (it's a 12MB model!)
Downloads last month
11
Safetensors
Model size
3.06M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for mascIT/bert-tiny-ita

Finetunes
1 model