distilbert/distilbert-base-multilingual-cased — Zero-Shot NLI Fine-Tune

Model Description

Fine-tuned distilbert/distilbert-base-multilingual-cased for zero-shot text classification via NLI.

Quick Start

from transformers import pipeline
clf = pipeline("zero-shot-classification", model="<your-hf-repo>")
result = clf("Your text here", candidate_labels=["label_a", "label_b"])
print(result)

Training Data

Dataset Label column
AyoubChLin/CompanyDocuments document_type
AyoubChLin/ARxiv_Metadata_50k labels
AyoubChLin/CNN_News_Articles_2011-2022 label

Training Details

Parameter Value
Base model distilbert/distilbert-base-multilingual-cased
Epochs 8
Learning rate 2e-05
Effective batch 128
Scheduler cosine_with_restarts
Precision bf16
Gradient Checkpointing False
LLRD True (factor=0.9)

Evaluation Results (Zero-Shot)

Dataset Acc@1 Macro-F1 MCC ROC-AUC Avg-Prec
CNN_News_Articles 0.9178 0.5244 0.8541 0.9671 0.7054
ARxiv_Metadata_top20 0.9575 0.1630 -0.0189 0.6084 0.1758

Framework Versions

  • transformers: 4.56.0
  • torch: 2.8.0+cu129
  • datasets: 4.8.5'
Downloads last month
9
Safetensors
Model size
0.1B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Datasets used to train AyoubChLin/distilbert-base-multilingual-cased-zeroshot-nli