YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

SIB200 Base Model

This model was trained on the SIB200 dataset using random data selection.

Training Parameters

  • Dataset: SIB200
  • Mode: Base
  • Selection Method: Random
  • Train Size: 700 examples
  • Epochs: 20
  • Batch Size: 8
  • Effective Batch Size: 32 (batch_size * gradient_accumulation_steps)
  • Learning Rate: 8e-06
  • Patience: 8
  • Max Length: 192
  • Gradient Accumulation Steps: 4
  • Warmup Ratio: 0.1
  • Weight Decay: 0.01
  • Optimizer: AdamW
  • Scheduler: cosine_with_warmup
  • Random Seed: 42

Performance

  • Overall Accuracy: 78.79%
  • Overall Loss: 0.0166

Language-Specific Performance

  • English (EN): 82.83%
  • German (DE): 87.88%
  • Arabic (AR): 54.55%
  • Spanish (ES): 87.88%
  • Hindi (HI): 80.81%
  • Swahili (SW): 78.79%

Model Information

  • Base Model: bert-base-multilingual-cased
  • Task: Topic Classification
  • Languages: 6 languages (EN, DE, AR, ES, HI, SW)
Downloads last month
5
Safetensors
Model size
0.2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including fledor/sib200_mbert_base_multilingual