henryscheible commited on
Commit
429c3ca
·
1 Parent(s): d9e01af

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -78
README.md CHANGED
@@ -2,8 +2,6 @@
2
  license: mit
3
  tags:
4
  - generated_from_trainer
5
- metrics:
6
- - accuracy
7
  model-index:
8
  - name: gpt2_winobias_classifieronly
9
  results: []
@@ -14,14 +12,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # gpt2_winobias_classifieronly
16
 
17
- This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 0.6987
20
- - Accuracy: 0.4924
21
- - Tp: 0.1660
22
- - Tn: 0.3264
23
- - Fp: 0.1736
24
- - Fn: 0.3340
25
 
26
  ## Model description
27
 
@@ -48,74 +39,6 @@ The following hyperparameters were used during training:
48
  - lr_scheduler_type: linear
49
  - num_epochs: 50
50
 
51
- ### Training results
52
-
53
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | Tp | Tn | Fp | Fn |
54
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:------:|:------:|:------:|
55
- | 0.7261 | 0.8 | 20 | 0.7092 | 0.5032 | 0.3428 | 0.1604 | 0.3396 | 0.1572 |
56
- | 0.7385 | 1.6 | 40 | 0.7073 | 0.5025 | 0.1673 | 0.3352 | 0.1648 | 0.3327 |
57
- | 0.7058 | 2.4 | 60 | 0.7054 | 0.5044 | 0.1970 | 0.3074 | 0.1926 | 0.3030 |
58
- | 0.7201 | 3.2 | 80 | 0.7048 | 0.5063 | 0.1963 | 0.3100 | 0.1900 | 0.3037 |
59
- | 0.7124 | 4.0 | 100 | 0.7040 | 0.5057 | 0.1976 | 0.3081 | 0.1919 | 0.3024 |
60
- | 0.7045 | 4.8 | 120 | 0.7046 | 0.4994 | 0.1534 | 0.3460 | 0.1540 | 0.3466 |
61
- | 0.725 | 5.6 | 140 | 0.7024 | 0.4918 | 0.2052 | 0.2866 | 0.2134 | 0.2948 |
62
- | 0.7225 | 6.4 | 160 | 0.7036 | 0.5025 | 0.1578 | 0.3447 | 0.1553 | 0.3422 |
63
- | 0.7037 | 7.2 | 180 | 0.7043 | 0.4975 | 0.1351 | 0.3624 | 0.1376 | 0.3649 |
64
- | 0.7313 | 8.0 | 200 | 0.7018 | 0.4962 | 0.1900 | 0.3062 | 0.1938 | 0.3100 |
65
- | 0.714 | 8.8 | 220 | 0.7023 | 0.5082 | 0.1730 | 0.3352 | 0.1648 | 0.3270 |
66
- | 0.7191 | 9.6 | 240 | 0.7018 | 0.5032 | 0.1787 | 0.3245 | 0.1755 | 0.3213 |
67
- | 0.7122 | 10.4 | 260 | 0.7013 | 0.5 | 0.1919 | 0.3081 | 0.1919 | 0.3081 |
68
- | 0.7147 | 11.2 | 280 | 0.7018 | 0.5006 | 0.1629 | 0.3378 | 0.1622 | 0.3371 |
69
- | 0.7238 | 12.0 | 300 | 0.7011 | 0.4987 | 0.1768 | 0.3220 | 0.1780 | 0.3232 |
70
- | 0.7146 | 12.8 | 320 | 0.7013 | 0.4962 | 0.1667 | 0.3295 | 0.1705 | 0.3333 |
71
- | 0.7108 | 13.6 | 340 | 0.7022 | 0.5019 | 0.1427 | 0.3592 | 0.1408 | 0.3573 |
72
- | 0.7213 | 14.4 | 360 | 0.6999 | 0.4937 | 0.2090 | 0.2847 | 0.2153 | 0.2910 |
73
- | 0.7202 | 15.2 | 380 | 0.7011 | 0.4962 | 0.1534 | 0.3428 | 0.1572 | 0.3466 |
74
- | 0.7136 | 16.0 | 400 | 0.7012 | 0.4981 | 0.1521 | 0.3460 | 0.1540 | 0.3479 |
75
- | 0.7204 | 16.8 | 420 | 0.7007 | 0.5 | 0.1641 | 0.3359 | 0.1641 | 0.3359 |
76
- | 0.7238 | 17.6 | 440 | 0.7004 | 0.4962 | 0.1673 | 0.3289 | 0.1711 | 0.3327 |
77
- | 0.7304 | 18.4 | 460 | 0.7002 | 0.4931 | 0.1705 | 0.3226 | 0.1774 | 0.3295 |
78
- | 0.7037 | 19.2 | 480 | 0.7003 | 0.4931 | 0.1648 | 0.3283 | 0.1717 | 0.3352 |
79
- | 0.7228 | 20.0 | 500 | 0.6998 | 0.4956 | 0.1818 | 0.3138 | 0.1862 | 0.3182 |
80
- | 0.7235 | 20.8 | 520 | 0.6997 | 0.4905 | 0.1761 | 0.3144 | 0.1856 | 0.3239 |
81
- | 0.7137 | 21.6 | 540 | 0.6993 | 0.4962 | 0.2014 | 0.2948 | 0.2052 | 0.2986 |
82
- | 0.7152 | 22.4 | 560 | 0.6993 | 0.4931 | 0.1963 | 0.2967 | 0.2033 | 0.3037 |
83
- | 0.7082 | 23.2 | 580 | 0.6992 | 0.4943 | 0.2014 | 0.2929 | 0.2071 | 0.2986 |
84
- | 0.7075 | 24.0 | 600 | 0.6995 | 0.4962 | 0.1774 | 0.3188 | 0.1812 | 0.3226 |
85
- | 0.7145 | 24.8 | 620 | 0.6991 | 0.4924 | 0.1976 | 0.2948 | 0.2052 | 0.3024 |
86
- | 0.7076 | 25.6 | 640 | 0.7007 | 0.4981 | 0.1332 | 0.3649 | 0.1351 | 0.3668 |
87
- | 0.7117 | 26.4 | 660 | 0.6998 | 0.4981 | 0.1597 | 0.3384 | 0.1616 | 0.3403 |
88
- | 0.696 | 27.2 | 680 | 0.6997 | 0.4987 | 0.1591 | 0.3396 | 0.1604 | 0.3409 |
89
- | 0.7176 | 28.0 | 700 | 0.6994 | 0.4981 | 0.1692 | 0.3289 | 0.1711 | 0.3308 |
90
- | 0.7148 | 28.8 | 720 | 0.6997 | 0.4981 | 0.1578 | 0.3403 | 0.1597 | 0.3422 |
91
- | 0.7147 | 29.6 | 740 | 0.6986 | 0.4981 | 0.2159 | 0.2822 | 0.2178 | 0.2841 |
92
- | 0.7281 | 30.4 | 760 | 0.6989 | 0.4956 | 0.1862 | 0.3093 | 0.1907 | 0.3138 |
93
- | 0.7052 | 31.2 | 780 | 0.6995 | 0.4981 | 0.1591 | 0.3390 | 0.1610 | 0.3409 |
94
- | 0.7301 | 32.0 | 800 | 0.6991 | 0.4975 | 0.1749 | 0.3226 | 0.1774 | 0.3251 |
95
- | 0.7222 | 32.8 | 820 | 0.6988 | 0.4962 | 0.1850 | 0.3112 | 0.1888 | 0.3150 |
96
- | 0.7202 | 33.6 | 840 | 0.6991 | 0.4949 | 0.1679 | 0.3270 | 0.1730 | 0.3321 |
97
- | 0.7104 | 34.4 | 860 | 0.6985 | 0.4937 | 0.2014 | 0.2923 | 0.2077 | 0.2986 |
98
- | 0.7207 | 35.2 | 880 | 0.6982 | 0.4937 | 0.2222 | 0.2715 | 0.2285 | 0.2778 |
99
- | 0.6985 | 36.0 | 900 | 0.6983 | 0.4981 | 0.2115 | 0.2866 | 0.2134 | 0.2885 |
100
- | 0.7073 | 36.8 | 920 | 0.6987 | 0.4975 | 0.1806 | 0.3169 | 0.1831 | 0.3194 |
101
- | 0.7186 | 37.6 | 940 | 0.6983 | 0.4981 | 0.2058 | 0.2923 | 0.2077 | 0.2942 |
102
- | 0.7084 | 38.4 | 960 | 0.6986 | 0.4956 | 0.1837 | 0.3119 | 0.1881 | 0.3163 |
103
- | 0.7197 | 39.2 | 980 | 0.6988 | 0.4987 | 0.1742 | 0.3245 | 0.1755 | 0.3258 |
104
- | 0.7137 | 40.0 | 1000 | 0.6987 | 0.4962 | 0.1761 | 0.3201 | 0.1799 | 0.3239 |
105
- | 0.7107 | 40.8 | 1020 | 0.6987 | 0.4968 | 0.1730 | 0.3239 | 0.1761 | 0.3270 |
106
- | 0.6927 | 41.6 | 1040 | 0.6988 | 0.4949 | 0.1692 | 0.3258 | 0.1742 | 0.3308 |
107
- | 0.7125 | 42.4 | 1060 | 0.6989 | 0.4937 | 0.1610 | 0.3327 | 0.1673 | 0.3390 |
108
- | 0.7038 | 43.2 | 1080 | 0.6990 | 0.4937 | 0.1572 | 0.3365 | 0.1635 | 0.3428 |
109
- | 0.7254 | 44.0 | 1100 | 0.6990 | 0.4943 | 0.1578 | 0.3365 | 0.1635 | 0.3422 |
110
- | 0.7122 | 44.8 | 1120 | 0.6990 | 0.4937 | 0.1559 | 0.3378 | 0.1622 | 0.3441 |
111
- | 0.7129 | 45.6 | 1140 | 0.6990 | 0.4937 | 0.1553 | 0.3384 | 0.1616 | 0.3447 |
112
- | 0.7204 | 46.4 | 1160 | 0.6988 | 0.4924 | 0.1635 | 0.3289 | 0.1711 | 0.3365 |
113
- | 0.716 | 47.2 | 1180 | 0.6987 | 0.4962 | 0.1717 | 0.3245 | 0.1755 | 0.3283 |
114
- | 0.7192 | 48.0 | 1200 | 0.6987 | 0.4956 | 0.1705 | 0.3251 | 0.1749 | 0.3295 |
115
- | 0.704 | 48.8 | 1220 | 0.6987 | 0.4956 | 0.1705 | 0.3251 | 0.1749 | 0.3295 |
116
- | 0.7072 | 49.6 | 1240 | 0.6987 | 0.4924 | 0.1660 | 0.3264 | 0.1736 | 0.3340 |
117
-
118
-
119
  ### Framework versions
120
 
121
  - Transformers 4.26.1
 
2
  license: mit
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: gpt2_winobias_classifieronly
7
  results: []
 
12
 
13
  # gpt2_winobias_classifieronly
14
 
15
+ This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
 
 
 
 
 
 
 
16
 
17
  ## Model description
18
 
 
39
  - lr_scheduler_type: linear
40
  - num_epochs: 50
41
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  ### Framework versions
43
 
44
  - Transformers 4.26.1