bitlabsdb commited on
Commit
e0fa989
·
verified ·
1 Parent(s): f834c7f

Upload trained model (Safetensors) Acc: 72.79%

Browse files
README.md CHANGED
@@ -4,7 +4,8 @@ tags:
4
  - fairsteer
5
  - bias-detection
6
  - tinyllama
7
- library_name: pytorch
 
8
  pipeline_tag: text-classification
9
  ---
10
 
@@ -13,11 +14,11 @@ pipeline_tag: text-classification
13
  Biased Activation Detection (BAD) classifier for TinyLlama-1.1B.
14
 
15
  ## Artifacts
16
- - **Model**: `pytorch_model.bin`
17
  - **Scaler**: `scaler.pkl` (StandardScaler)
18
  - **Config**: `config.json`
19
 
20
  ## Stats
21
- - **Balanced Accuracy**: 76.47%
22
- - **Best Layer**: 11
23
- - **Training Date**: 2025-11-22
 
4
  - fairsteer
5
  - bias-detection
6
  - tinyllama
7
+ - safetensors
8
+ library_name: safetensors
9
  pipeline_tag: text-classification
10
  ---
11
 
 
14
  Biased Activation Detection (BAD) classifier for TinyLlama-1.1B.
15
 
16
  ## Artifacts
17
+ - **Model**: `model.safetensors` (SafeTensors format)
18
  - **Scaler**: `scaler.pkl` (StandardScaler)
19
  - **Config**: `config.json`
20
 
21
  ## Stats
22
+ - **Balanced Accuracy**: 72.79%
23
+ - **Best Layer**: 20
24
+ - **Training Date**: 2025-12-12
config.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "input_dim": 2048,
3
- "layer_idx": 11,
4
  "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
5
- "best_val_bal_acc": 0.7647058823529411,
6
  "training_method": "GroupSplit + Standardized + Balanced",
7
- "training_date": "2025-11-22T06:49:56.551085"
8
  }
 
1
  {
2
  "input_dim": 2048,
3
+ "layer_idx": 20,
4
  "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
5
+ "best_val_bal_acc": 0.7279477947794779,
6
  "training_method": "GroupSplit + Standardized + Balanced",
7
+ "training_date": "2025-12-12T05:40:01.622551"
8
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9adaca3860a6cb9885b0c51784436317c98fc0998c0f25258ba1a9cb30de66a
3
+ size 8348
scaler.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee05e1401722adc64f8dcec1129440add2e2b42b51e5b16de6c49bb7026287cf
3
  size 49615
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1e2def07ab40c3436e01391149a252fb4634cdf7d6b89b5ae3b29c606783586
3
  size 49615
training_dashboard.png CHANGED

Git LFS Details

  • SHA256: 3479d31c43a2fb5b90d9f7cbaa6cf5167c828dfe5b2594c0d5db3124accbfcb9
  • Pointer size: 131 Bytes
  • Size of remote file: 146 kB

Git LFS Details

  • SHA256: 5295da7a9430b032977ca2fd5dd4fbe53932fb23bf40d516fd0d58e360a3380b
  • Pointer size: 131 Bytes
  • Size of remote file: 147 kB
training_history.json CHANGED
@@ -1,62 +1,98 @@
1
  {
2
  "val_auc": [
3
- 0.6721453287197232,
4
- 0.7370242214532872,
5
- 0.7352941176470589,
6
- 0.7448096885813149,
7
- 0.7543252595155708,
8
- 0.7569204152249136,
9
- 0.7577854671280276,
10
- 0.759515570934256,
11
- 0.7655709342560553,
12
- 0.7698961937716262,
13
- 0.7707612456747406,
14
- 0.7716262975778547,
15
- 0.7707612456747405,
16
- 0.7690311418685121,
17
- 0.7698961937716263,
18
- 0.7690311418685121,
19
- 0.7690311418685121,
20
- 0.7681660899653979
 
 
 
 
 
 
 
 
 
 
 
 
21
  ],
22
  "val_bal_acc": [
23
- 0.6323529411764706,
24
- 0.7058823529411764,
25
- 0.7647058823529411,
26
- 0.7647058823529411,
27
- 0.7352941176470589,
28
- 0.7205882352941176,
29
- 0.7205882352941176,
30
- 0.7205882352941176,
31
- 0.7205882352941176,
32
- 0.7205882352941176,
33
- 0.7205882352941176,
34
- 0.7205882352941176,
35
- 0.7205882352941176,
36
- 0.7205882352941176,
37
- 0.7205882352941176,
38
- 0.7205882352941176,
39
- 0.7205882352941176,
40
- 0.7205882352941176
 
 
 
 
 
 
 
 
 
 
 
 
41
  ],
42
  "train_loss": [
43
- 0.7645134925842285,
44
- 0.6531355579694113,
45
- 0.5934327244758606,
46
- 0.5382237235705057,
47
- 0.5212106804052988,
48
- 0.5007542173067728,
49
- 0.49714147051175434,
50
- 0.48595739404360455,
51
- 0.48224801818529767,
52
- 0.4633047580718994,
53
- 0.4640958805878957,
54
- 0.4591399331887563,
55
- 0.4668729603290558,
56
- 0.45914029081662494,
57
- 0.4610348045825958,
58
- 0.46088655789693195,
59
- 0.45236412684122723,
60
- 0.46567730108896893
 
 
 
 
 
 
 
 
 
 
 
 
61
  ]
62
  }
 
1
  {
2
  "val_auc": [
3
+ 0.7273940490358668,
4
+ 0.7504346519260386,
5
+ 0.7575085816412425,
6
+ 0.7573322701307035,
7
+ 0.7596420420619919,
8
+ 0.7644952163983275,
9
+ 0.7669964903781107,
10
+ 0.7635729482039112,
11
+ 0.7664611758205524,
12
+ 0.7653390001466394,
13
+ 0.7669162843377048,
14
+ 0.7698281079233036,
15
+ 0.7678008799979907,
16
+ 0.7703854993960242,
17
+ 0.7706289530843273,
18
+ 0.7703307121891307,
19
+ 0.7706585239982144,
20
+ 0.7678606294472832,
21
+ 0.7682736095391718,
22
+ 0.7705408479641844,
23
+ 0.7703784104783117,
24
+ 0.7698073475214309,
25
+ 0.767579604495103,
26
+ 0.7694949287908088,
27
+ 0.7691457489583342,
28
+ 0.7687167681664656,
29
+ 0.7699388975801271,
30
+ 0.7697089128354779,
31
+ 0.7699554046313721,
32
+ 0.7698038030625745
33
  ],
34
  "val_bal_acc": [
35
+ 0.6671917191719172,
36
+ 0.6874437443744374,
37
+ 0.7031953195319531,
38
+ 0.7029702970297029,
39
+ 0.7088208820882088,
40
+ 0.7200720072007201,
41
+ 0.711971197119712,
42
+ 0.711971197119712,
43
+ 0.716021602160216,
44
+ 0.7133213321332134,
45
+ 0.7223222322232223,
46
+ 0.7236723672367237,
47
+ 0.7209720972097209,
48
+ 0.725922592259226,
49
+ 0.7279477947794779,
50
+ 0.7232223222322232,
51
+ 0.7214221422142215,
52
+ 0.718046804680468,
53
+ 0.7178217821782178,
54
+ 0.7205220522052205,
55
+ 0.7225472547254725,
56
+ 0.7184968496849685,
57
+ 0.7198469846984699,
58
+ 0.7227722772277227,
59
+ 0.7193969396939695,
60
+ 0.7252475247524752,
61
+ 0.7216471647164716,
62
+ 0.7227722772277227,
63
+ 0.7214221422142214,
64
+ 0.7241224122412241
65
  ],
66
  "train_loss": [
67
+ 0.6589798988967106,
68
+ 0.5834655373260893,
69
+ 0.5626362512851584,
70
+ 0.5517601516740075,
71
+ 0.5468299312838193,
72
+ 0.5445838414389511,
73
+ 0.5378933320785391,
74
+ 0.5344717323780059,
75
+ 0.5338149806548809,
76
+ 0.5299872453870444,
77
+ 0.5244478642940521,
78
+ 0.5223940512229657,
79
+ 0.5225685875991295,
80
+ 0.5207774322608422,
81
+ 0.5200736000620085,
82
+ 0.5207284010689834,
83
+ 0.5184559558999949,
84
+ 0.517919123994893,
85
+ 0.518367536725669,
86
+ 0.5166512495484845,
87
+ 0.5150178672938511,
88
+ 0.5143926667756048,
89
+ 0.5142317825350268,
90
+ 0.5123990134946231,
91
+ 0.5112694777291397,
92
+ 0.513259904754573,
93
+ 0.512855320757833,
94
+ 0.5109211087226868,
95
+ 0.510224807467954,
96
+ 0.5088399624002391
97
  ]
98
  }