DornierDo17 commited on
Commit
2b36b15
·
verified ·
1 Parent(s): 3f2b802

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -3
README.md CHANGED
@@ -1,3 +1,31 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - stanfordnlp/sst2
5
+ language:
6
+ - en
7
+ metrics:
8
+ - accuracy
9
+ base_model:
10
+ - DornierDo17/RoBERTa_17.7M
11
+ pipeline_tag: text-classification
12
+ library_name: transformers
13
+ tags:
14
+ - '#sst2'
15
+ - '#text-classification'
16
+ - '#ai'
17
+ ---
18
+
19
+
20
+
21
+ This model is a fine-tuned version of my base MiniRoBERTa (17.7M parameters) model. The goal of this fine-tuning experiment was to demonstrate that a RoBERTa-style model, built entirely from scratch and trained on a single GPU with limited compute, can still learn meaningful patterns and adapt effectively to downstream tasks.
22
+
23
+ The model was fine-tuned on the SST-2 sentiment classification dataset and achieved an accuracy of 80%, which is a strong result given the scale and simplicity of the pretraining setup.
24
+
25
+ This validates that the model has learned generalizable representations and can be adapted successfully to real-world NLP tasks through fine-tuning.
26
+
27
+ #### Highlights:
28
+ - Fine-tuned from scratch-trained RoBERTa (17.7M) model
29
+ - Dataset: SST-2 (Stanford Sentiment Treebank)
30
+ - Accuracy: **80%**
31
+ - Trained on: Single GPU (limited compute)