File size: 2,027 Bytes
73539b6 1f16026 dc4329c 1f16026 73539b6 1f16026 73539b6 dc4329c 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 d40e34a 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 d40e34a 73539b6 1f16026 73539b6 dc4329c 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 d40e34a dc4329c 73539b6 1f16026 73539b6 dc4329c 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 73539b6 1f16026 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 |
---
library_name: transformers
tags:
- tone
datasets:
- Dc-4nderson/tone_dataset2
language:
- en
metrics:
- accuracy
- f1
base_model:
- distilbert/distilbert-base-uncased
pipeline_tag: text-classification
---
DistilBERT Tone Classification Model
This model fine-tunes distilbert-base-uncased to classify tone into 7 categories relevant to community and mentorship transcripts.
π Labels
uplifting
thoughtful
practical
reflective
motivational
informative
optimistic
π Dataset
The model is trained on the tone-dataset
,
a dataset containing 1000+ labeled examples created for the MyVillageProject tone classification task.
Data includes first-person and third-person statements, anecdotes, factual notes, and reflective entries.
π Training
Base model: distilbert-base-uncased
Optimizer: AdamW (lr=2e-5)
Batch size: 16
Epochs: 8
Loss: CrossEntropy
Metrics: Accuracy + Weighted F1
π Validation Metrics
Epoch Training Loss Validation Loss Accuracy F1
1 No log 1.281651 0.782288 0.778880
2 No log 0.779447 0.845018 0.843397
3 No log 0.566092 0.859779 0.856186
4 No log 0.415437 0.892989 0.892445
5 No log 0.340598 0.915129 0.914765
6 0.729500 0.307513 0.922509 0.922262
7 0.729500 0.296827 0.915129 0.915210
8 0.729500 0.285301 0.922509 0.922262
Final Training Summary:
TrainOutput(global_step=704,
training_loss=0.5666945034807379,
metrics={'train_runtime': 42.6317,
'train_samples_per_second': 261.402,
'train_steps_per_second': 16.514,
'total_flos': 369087080441856.0,
'train_loss': 0.5666945034807379,
'epoch': 8.0})
π» Usage
from transformers import pipeline
classifier = pipeline("text-classification", model="Dc-4nderson/tone-distilbert")
text = "Ronnie mentioned the turnout was twice what they expected, and it felt like a victory."
print(classifier(text))
Output:
[{'label': 'uplifting'}]
π₯ Maintainer
Dequan Anderson/ Dc-4nderson
|