Upload trained model (Safetensors) Acc: 72.79%
Browse files- README.md +6 -5
- config.json +3 -3
- model.safetensors +3 -0
- scaler.pkl +1 -1
- training_dashboard.png +2 -2
- training_history.json +90 -54
README.md
CHANGED
|
@@ -4,7 +4,8 @@ tags:
|
|
| 4 |
- fairsteer
|
| 5 |
- bias-detection
|
| 6 |
- tinyllama
|
| 7 |
-
|
|
|
|
| 8 |
pipeline_tag: text-classification
|
| 9 |
---
|
| 10 |
|
|
@@ -13,11 +14,11 @@ pipeline_tag: text-classification
|
|
| 13 |
Biased Activation Detection (BAD) classifier for TinyLlama-1.1B.
|
| 14 |
|
| 15 |
## Artifacts
|
| 16 |
-
- **Model**: `
|
| 17 |
- **Scaler**: `scaler.pkl` (StandardScaler)
|
| 18 |
- **Config**: `config.json`
|
| 19 |
|
| 20 |
## Stats
|
| 21 |
-
- **Balanced Accuracy**:
|
| 22 |
-
- **Best Layer**:
|
| 23 |
-
- **Training Date**: 2025-
|
|
|
|
| 4 |
- fairsteer
|
| 5 |
- bias-detection
|
| 6 |
- tinyllama
|
| 7 |
+
- safetensors
|
| 8 |
+
library_name: safetensors
|
| 9 |
pipeline_tag: text-classification
|
| 10 |
---
|
| 11 |
|
|
|
|
| 14 |
Biased Activation Detection (BAD) classifier for TinyLlama-1.1B.
|
| 15 |
|
| 16 |
## Artifacts
|
| 17 |
+
- **Model**: `model.safetensors` (SafeTensors format)
|
| 18 |
- **Scaler**: `scaler.pkl` (StandardScaler)
|
| 19 |
- **Config**: `config.json`
|
| 20 |
|
| 21 |
## Stats
|
| 22 |
+
- **Balanced Accuracy**: 72.79%
|
| 23 |
+
- **Best Layer**: 20
|
| 24 |
+
- **Training Date**: 2025-12-12
|
config.json
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"input_dim": 2048,
|
| 3 |
-
"layer_idx":
|
| 4 |
"base_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
| 5 |
-
"best_val_bal_acc": 0.
|
| 6 |
"training_method": "GroupSplit + Standardized + Balanced",
|
| 7 |
-
"training_date": "2025-
|
| 8 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"input_dim": 2048,
|
| 3 |
+
"layer_idx": 20,
|
| 4 |
"base_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
| 5 |
+
"best_val_bal_acc": 0.7279477947794779,
|
| 6 |
"training_method": "GroupSplit + Standardized + Balanced",
|
| 7 |
+
"training_date": "2025-12-12T05:40:01.622551"
|
| 8 |
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e9adaca3860a6cb9885b0c51784436317c98fc0998c0f25258ba1a9cb30de66a
|
| 3 |
+
size 8348
|
scaler.pkl
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 49615
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d1e2def07ab40c3436e01391149a252fb4634cdf7d6b89b5ae3b29c606783586
|
| 3 |
size 49615
|
training_dashboard.png
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
training_history.json
CHANGED
|
@@ -1,62 +1,98 @@
|
|
| 1 |
{
|
| 2 |
"val_auc": [
|
| 3 |
-
0.
|
| 4 |
-
0.
|
| 5 |
-
0.
|
| 6 |
-
0.
|
| 7 |
-
0.
|
| 8 |
-
0.
|
| 9 |
-
0.
|
| 10 |
-
0.
|
| 11 |
-
0.
|
| 12 |
-
0.
|
| 13 |
-
0.
|
| 14 |
-
0.
|
| 15 |
-
0.
|
| 16 |
-
0.
|
| 17 |
-
0.
|
| 18 |
-
0.
|
| 19 |
-
0.
|
| 20 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
],
|
| 22 |
"val_bal_acc": [
|
| 23 |
-
0.
|
| 24 |
-
0.
|
| 25 |
-
0.
|
| 26 |
-
0.
|
| 27 |
-
0.
|
| 28 |
-
0.
|
| 29 |
-
0.
|
| 30 |
-
0.
|
| 31 |
-
0.
|
| 32 |
-
0.
|
| 33 |
-
0.
|
| 34 |
-
0.
|
| 35 |
-
0.
|
| 36 |
-
0.
|
| 37 |
-
0.
|
| 38 |
-
0.
|
| 39 |
-
0.
|
| 40 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 41 |
],
|
| 42 |
"train_loss": [
|
| 43 |
-
0.
|
| 44 |
-
0.
|
| 45 |
-
0.
|
| 46 |
-
0.
|
| 47 |
-
0.
|
| 48 |
-
0.
|
| 49 |
-
0.
|
| 50 |
-
0.
|
| 51 |
-
0.
|
| 52 |
-
0.
|
| 53 |
-
0.
|
| 54 |
-
0.
|
| 55 |
-
0.
|
| 56 |
-
0.
|
| 57 |
-
0.
|
| 58 |
-
0.
|
| 59 |
-
0.
|
| 60 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 61 |
]
|
| 62 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"val_auc": [
|
| 3 |
+
0.7273940490358668,
|
| 4 |
+
0.7504346519260386,
|
| 5 |
+
0.7575085816412425,
|
| 6 |
+
0.7573322701307035,
|
| 7 |
+
0.7596420420619919,
|
| 8 |
+
0.7644952163983275,
|
| 9 |
+
0.7669964903781107,
|
| 10 |
+
0.7635729482039112,
|
| 11 |
+
0.7664611758205524,
|
| 12 |
+
0.7653390001466394,
|
| 13 |
+
0.7669162843377048,
|
| 14 |
+
0.7698281079233036,
|
| 15 |
+
0.7678008799979907,
|
| 16 |
+
0.7703854993960242,
|
| 17 |
+
0.7706289530843273,
|
| 18 |
+
0.7703307121891307,
|
| 19 |
+
0.7706585239982144,
|
| 20 |
+
0.7678606294472832,
|
| 21 |
+
0.7682736095391718,
|
| 22 |
+
0.7705408479641844,
|
| 23 |
+
0.7703784104783117,
|
| 24 |
+
0.7698073475214309,
|
| 25 |
+
0.767579604495103,
|
| 26 |
+
0.7694949287908088,
|
| 27 |
+
0.7691457489583342,
|
| 28 |
+
0.7687167681664656,
|
| 29 |
+
0.7699388975801271,
|
| 30 |
+
0.7697089128354779,
|
| 31 |
+
0.7699554046313721,
|
| 32 |
+
0.7698038030625745
|
| 33 |
],
|
| 34 |
"val_bal_acc": [
|
| 35 |
+
0.6671917191719172,
|
| 36 |
+
0.6874437443744374,
|
| 37 |
+
0.7031953195319531,
|
| 38 |
+
0.7029702970297029,
|
| 39 |
+
0.7088208820882088,
|
| 40 |
+
0.7200720072007201,
|
| 41 |
+
0.711971197119712,
|
| 42 |
+
0.711971197119712,
|
| 43 |
+
0.716021602160216,
|
| 44 |
+
0.7133213321332134,
|
| 45 |
+
0.7223222322232223,
|
| 46 |
+
0.7236723672367237,
|
| 47 |
+
0.7209720972097209,
|
| 48 |
+
0.725922592259226,
|
| 49 |
+
0.7279477947794779,
|
| 50 |
+
0.7232223222322232,
|
| 51 |
+
0.7214221422142215,
|
| 52 |
+
0.718046804680468,
|
| 53 |
+
0.7178217821782178,
|
| 54 |
+
0.7205220522052205,
|
| 55 |
+
0.7225472547254725,
|
| 56 |
+
0.7184968496849685,
|
| 57 |
+
0.7198469846984699,
|
| 58 |
+
0.7227722772277227,
|
| 59 |
+
0.7193969396939695,
|
| 60 |
+
0.7252475247524752,
|
| 61 |
+
0.7216471647164716,
|
| 62 |
+
0.7227722772277227,
|
| 63 |
+
0.7214221422142214,
|
| 64 |
+
0.7241224122412241
|
| 65 |
],
|
| 66 |
"train_loss": [
|
| 67 |
+
0.6589798988967106,
|
| 68 |
+
0.5834655373260893,
|
| 69 |
+
0.5626362512851584,
|
| 70 |
+
0.5517601516740075,
|
| 71 |
+
0.5468299312838193,
|
| 72 |
+
0.5445838414389511,
|
| 73 |
+
0.5378933320785391,
|
| 74 |
+
0.5344717323780059,
|
| 75 |
+
0.5338149806548809,
|
| 76 |
+
0.5299872453870444,
|
| 77 |
+
0.5244478642940521,
|
| 78 |
+
0.5223940512229657,
|
| 79 |
+
0.5225685875991295,
|
| 80 |
+
0.5207774322608422,
|
| 81 |
+
0.5200736000620085,
|
| 82 |
+
0.5207284010689834,
|
| 83 |
+
0.5184559558999949,
|
| 84 |
+
0.517919123994893,
|
| 85 |
+
0.518367536725669,
|
| 86 |
+
0.5166512495484845,
|
| 87 |
+
0.5150178672938511,
|
| 88 |
+
0.5143926667756048,
|
| 89 |
+
0.5142317825350268,
|
| 90 |
+
0.5123990134946231,
|
| 91 |
+
0.5112694777291397,
|
| 92 |
+
0.513259904754573,
|
| 93 |
+
0.512855320757833,
|
| 94 |
+
0.5109211087226868,
|
| 95 |
+
0.510224807467954,
|
| 96 |
+
0.5088399624002391
|
| 97 |
]
|
| 98 |
}
|