Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,31 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
datasets:
|
| 4 |
+
- stanfordnlp/sst2
|
| 5 |
+
language:
|
| 6 |
+
- en
|
| 7 |
+
metrics:
|
| 8 |
+
- accuracy
|
| 9 |
+
base_model:
|
| 10 |
+
- DornierDo17/RoBERTa_17.7M
|
| 11 |
+
pipeline_tag: text-classification
|
| 12 |
+
library_name: transformers
|
| 13 |
+
tags:
|
| 14 |
+
- '#sst2'
|
| 15 |
+
- '#text-classification'
|
| 16 |
+
- '#ai'
|
| 17 |
+
---
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
This model is a fine-tuned version of my base MiniRoBERTa (17.7M parameters) model. The goal of this fine-tuning experiment was to demonstrate that a RoBERTa-style model, built entirely from scratch and trained on a single GPU with limited compute, can still learn meaningful patterns and adapt effectively to downstream tasks.
|
| 22 |
+
|
| 23 |
+
The model was fine-tuned on the SST-2 sentiment classification dataset and achieved an accuracy of 80%, which is a strong result given the scale and simplicity of the pretraining setup.
|
| 24 |
+
|
| 25 |
+
This validates that the model has learned generalizable representations and can be adapted successfully to real-world NLP tasks through fine-tuning.
|
| 26 |
+
|
| 27 |
+
#### Highlights:
|
| 28 |
+
- Fine-tuned from scratch-trained RoBERTa (17.7M) model
|
| 29 |
+
- Dataset: SST-2 (Stanford Sentiment Treebank)
|
| 30 |
+
- Accuracy: **80%**
|
| 31 |
+
- Trained on: Single GPU (limited compute)
|