YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Argument–Keypoint Matching with DistilBERT

This model predicts whether an argument is correctly matched (Apparié) or not (Non-Apparié) with a given key point.

Model Description

  • Base Model: DistilBERT (uncased)
  • Task: Binary text-pair classification
  • Training Data: IBM ArgKP-2023 dataset (~32,000 examples)
  • Labels:
    • 0 — Non-Apparié
    • 1 — Apparié
  • Input: (argument, key_point)
  • Output: Predicted class + probabilities

Performance

  • Strong accuracy and F1 score on evaluation data
  • Reliable predictions across both labels

Training

Trained on a balanced argument–keypoint dataset

Exported using save_pretrained

Citation

@misc{argument-keypoint-matching,
  author       = {Malek Messaoudi},
  title        = {Argument–Keypoint Matching with DistilBERT},
  year         = {2025},
  publisher    = {Hugging Face},
  howpublished = {{\\url{{https://huggingface.co/NLP-Debater-Project/destlibert-keypoint-matching}}}}
}

License

MIT License

Downloads last month
67
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Space using NLP-Debater-Project/distilBert-keypoint-matching 1