mletsana commited on
Commit
8057a1e
·
verified ·
1 Parent(s): 464cbae

Upload test_metrics.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. test_metrics.json +43 -0
test_metrics.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "accuracy": 0.7323869610935857,
3
+ "macro_f1": 0.548414196565388,
4
+ "report": {
5
+ "0": {
6
+ "precision": 0.7825503355704698,
7
+ "recall": 0.7333333333333333,
8
+ "f1-score": 0.7571428571428571,
9
+ "support": 795.0
10
+ },
11
+ "1": {
12
+ "precision": 0.7420435510887772,
13
+ "recall": 0.714516129032258,
14
+ "f1-score": 0.7280197206244864,
15
+ "support": 620.0
16
+ },
17
+ "2": {
18
+ "precision": 0.0,
19
+ "recall": 0.0,
20
+ "f1-score": 0.0,
21
+ "support": 11.0
22
+ },
23
+ "3": {
24
+ "precision": 0.6553571428571429,
25
+ "recall": 0.7710084033613446,
26
+ "f1-score": 0.7084942084942085,
27
+ "support": 476.0
28
+ },
29
+ "accuracy": 0.7323869610935857,
30
+ "macro avg": {
31
+ "precision": 0.5449877573790974,
32
+ "recall": 0.5547144664317339,
33
+ "f1-score": 0.548414196565388,
34
+ "support": 1902.0
35
+ },
36
+ "weighted avg": {
37
+ "precision": 0.7329887058115485,
38
+ "recall": 0.7323869610935857,
39
+ "f1-score": 0.7310957105462651,
40
+ "support": 1902.0
41
+ }
42
+ }
43
+ }