DominicTWHV commited on
Commit
d8aa5ee
·
verified ·
1 Parent(s): bc99883

Added raw eval metrics in JSON form

Browse files
Files changed (1) hide show
  1. README.md +64 -0
README.md CHANGED
@@ -54,6 +54,38 @@ LABEL_THRESHOLDS = {
54
  }
55
  ```
56
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  ![Recall Metrics](assets/graphs/untuned/recall_deberta.png)
58
  ![Precision Metrics](assets/graphs/untuned/precision_deberta.png)
59
  ![F1 Metrics](assets/graphs/untuned/f1_deberta.png)
@@ -75,6 +107,38 @@ LABEL_THRESHOLDS = {
75
  }
76
  ```
77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
  ![Recall Metrics](assets/graphs/tuned/recall_deberta.png)
79
  ![Precision Metrics](assets/graphs/tuned/precision_deberta.png)
80
  ![F1 Metrics](assets/graphs/tuned/f1_deberta.png)
 
54
  }
55
  ```
56
 
57
+ **Raw Eval Metrics:**
58
+
59
+ ```json
60
+ {
61
+ "eval_loss":0.16034406423568726,
62
+ "eval_precision":0.6059971310039647,
63
+ "eval_recall":0.9138250950483955,
64
+ "eval_f1":0.7164361696270752,
65
+ "eval_precision_scam":0.9117559964465501,
66
+ "eval_recall_scam":0.9532507739938081,
67
+ "eval_f1_scam":0.9320417738761919,
68
+ "eval_precision_violence":0.42734150795721365,
69
+ "eval_recall_violence":0.8970427163198248,
70
+ "eval_f1_violence":0.5789008658773634,
71
+ "eval_precision_harassment":0.7726063829787234,
72
+ "eval_recall_harassment":0.9423076923076923,
73
+ "eval_f1_harassment":0.8490605427974948,
74
+ "eval_precision_hate_speech":0.429821819318537,
75
+ "eval_recall_hate_speech":0.8969341161121983,
76
+ "eval_f1_hate_speech":0.5811496196111581,
77
+ "eval_precision_toxicity":0.5737432488574989,
78
+ "eval_recall_toxicity":0.8712933753943217,
79
+ "eval_f1_toxicity":0.6918837675350702,
80
+ "eval_precision_obscenity":0.5207138304652645,
81
+ "eval_recall_obscenity":0.9221218961625283,
82
+ "eval_f1_obscenity":0.6655804480651731,
83
+ "eval_runtime":247.1414,
84
+ "eval_samples_per_second":117.512,
85
+ "eval_steps_per_second":2.452
86
+ }
87
+ ```
88
+
89
  ![Recall Metrics](assets/graphs/untuned/recall_deberta.png)
90
  ![Precision Metrics](assets/graphs/untuned/precision_deberta.png)
91
  ![F1 Metrics](assets/graphs/untuned/f1_deberta.png)
 
107
  }
108
  ```
109
 
110
+ **Raw Eval Metrics:**
111
+
112
+ ```json
113
+ {
114
+ "eval_loss":0.16034406423568726,
115
+ "eval_precision":0.6939850223558622,
116
+ "eval_recall":0.8150767410772812,
117
+ "eval_f1":0.7475019013835578,
118
+ "eval_precision_scam":0.9255447941888619,
119
+ "eval_recall_scam":0.9467492260061919,
120
+ "eval_f1_scam":0.936026936026936,
121
+ "eval_precision_violence":0.5140955364134691,
122
+ "eval_recall_violence":0.7190580503833516,
123
+ "eval_f1_violence":0.5995433789954338,
124
+ "eval_precision_harassment":0.8238218763510592,
125
+ "eval_recall_harassment":0.8829935125115848,
126
+ "eval_f1_harassment":0.8523820174457616,
127
+ "eval_precision_hate_speech":0.5606936416184971,
128
+ "eval_recall_hate_speech":0.6960208741030659,
129
+ "eval_f1_hate_speech":0.6210710128055879,
130
+ "eval_precision_toxicity":0.6890574214517876,
131
+ "eval_recall_toxicity":0.8025236593059937,
132
+ "eval_f1_toxicity":0.7414747886913436,
133
+ "eval_precision_obscenity":0.6506968641114983,
134
+ "eval_recall_obscenity":0.8431151241534989,
135
+ "eval_f1_obscenity":0.7345132743362832,
136
+ "eval_runtime":378.4334,
137
+ "eval_samples_per_second":76.743,
138
+ "eval_steps_per_second":1.601
139
+ }
140
+ ```
141
+
142
  ![Recall Metrics](assets/graphs/tuned/recall_deberta.png)
143
  ![Precision Metrics](assets/graphs/tuned/precision_deberta.png)
144
  ![F1 Metrics](assets/graphs/tuned/f1_deberta.png)