gokul-pv's picture
Update README.md
f0e8a0e verified
---
library_name: transformers
language:
- en
base_model:
- microsoft/deberta-base
pipeline_tag: text-classification
---
# Model Card
This model is a finetuned version of [microsoft/deberta-base](https://huggingface.co/microsoft/deberta-base) on the [Onion or Not](https://www.kaggle.com/datasets/chrisfilo/onion-or-not) dataset.
The model was fine-tuned for 5 epochs with a learning rate of 2e-5 and a linear schedule. Random token dropout was implemented during training to avoid overfitting.
The classification report is shown below:
```
Final Validation Accuracy: 93.31%
Final Classification Report:
precision recall f1-score support
NotOnion 0.94 0.96 0.95 3000
Onion 0.93 0.89 0.91 1800
accuracy 0.93 4800
macro avg 0.93 0.92 0.93 4800
weighted avg 0.93 0.93 0.93 4800
```
Running inference on a new sample gave the correct prediction:
```
Running example inference...
Text: Man With Fogged-Up Glasses Forced To Finish Soup Using Other Senses
Prediction: Onion
Confidence: 87.76%
Probabilities:
NotOnion: 12.24%
Onion: 87.76%
```