legacy-datasets/mc4
Updated • 1.76k • 153
How to use rasyosef/bert-mini-amharic-16k with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("fill-mask", model="rasyosef/bert-mini-amharic-16k") # Load model directly
from transformers import AutoTokenizer, AutoModelForMaskedLM
tokenizer = AutoTokenizer.from_pretrained("rasyosef/bert-mini-amharic-16k")
model = AutoModelForMaskedLM.from_pretrained("rasyosef/bert-mini-amharic-16k")This model has the same architecture as bert-mini and was pretrained from scratch using the Amharic subsets of the oscar and mc4 datasets, on a total of 165 Million tokens.
It achieves the following results on the evaluation set:
Loss: 2.59Perplexity: 13.33Even though this model only has 7.5 Million parameters, its perplexity score is comparable to the 36x larger 279 Million parameter xlm-roberta-base model on the same Amharic evaluation set.