Commit ·
429c3ca
1
Parent(s): d9e01af
update model card README.md
Browse files
README.md
CHANGED
|
@@ -2,8 +2,6 @@
|
|
| 2 |
license: mit
|
| 3 |
tags:
|
| 4 |
- generated_from_trainer
|
| 5 |
-
metrics:
|
| 6 |
-
- accuracy
|
| 7 |
model-index:
|
| 8 |
- name: gpt2_winobias_classifieronly
|
| 9 |
results: []
|
|
@@ -14,14 +12,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 14 |
|
| 15 |
# gpt2_winobias_classifieronly
|
| 16 |
|
| 17 |
-
This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on
|
| 18 |
-
It achieves the following results on the evaluation set:
|
| 19 |
-
- Loss: 0.6987
|
| 20 |
-
- Accuracy: 0.4924
|
| 21 |
-
- Tp: 0.1660
|
| 22 |
-
- Tn: 0.3264
|
| 23 |
-
- Fp: 0.1736
|
| 24 |
-
- Fn: 0.3340
|
| 25 |
|
| 26 |
## Model description
|
| 27 |
|
|
@@ -48,74 +39,6 @@ The following hyperparameters were used during training:
|
|
| 48 |
- lr_scheduler_type: linear
|
| 49 |
- num_epochs: 50
|
| 50 |
|
| 51 |
-
### Training results
|
| 52 |
-
|
| 53 |
-
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Tp | Tn | Fp | Fn |
|
| 54 |
-
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:------:|:------:|:------:|
|
| 55 |
-
| 0.7261 | 0.8 | 20 | 0.7092 | 0.5032 | 0.3428 | 0.1604 | 0.3396 | 0.1572 |
|
| 56 |
-
| 0.7385 | 1.6 | 40 | 0.7073 | 0.5025 | 0.1673 | 0.3352 | 0.1648 | 0.3327 |
|
| 57 |
-
| 0.7058 | 2.4 | 60 | 0.7054 | 0.5044 | 0.1970 | 0.3074 | 0.1926 | 0.3030 |
|
| 58 |
-
| 0.7201 | 3.2 | 80 | 0.7048 | 0.5063 | 0.1963 | 0.3100 | 0.1900 | 0.3037 |
|
| 59 |
-
| 0.7124 | 4.0 | 100 | 0.7040 | 0.5057 | 0.1976 | 0.3081 | 0.1919 | 0.3024 |
|
| 60 |
-
| 0.7045 | 4.8 | 120 | 0.7046 | 0.4994 | 0.1534 | 0.3460 | 0.1540 | 0.3466 |
|
| 61 |
-
| 0.725 | 5.6 | 140 | 0.7024 | 0.4918 | 0.2052 | 0.2866 | 0.2134 | 0.2948 |
|
| 62 |
-
| 0.7225 | 6.4 | 160 | 0.7036 | 0.5025 | 0.1578 | 0.3447 | 0.1553 | 0.3422 |
|
| 63 |
-
| 0.7037 | 7.2 | 180 | 0.7043 | 0.4975 | 0.1351 | 0.3624 | 0.1376 | 0.3649 |
|
| 64 |
-
| 0.7313 | 8.0 | 200 | 0.7018 | 0.4962 | 0.1900 | 0.3062 | 0.1938 | 0.3100 |
|
| 65 |
-
| 0.714 | 8.8 | 220 | 0.7023 | 0.5082 | 0.1730 | 0.3352 | 0.1648 | 0.3270 |
|
| 66 |
-
| 0.7191 | 9.6 | 240 | 0.7018 | 0.5032 | 0.1787 | 0.3245 | 0.1755 | 0.3213 |
|
| 67 |
-
| 0.7122 | 10.4 | 260 | 0.7013 | 0.5 | 0.1919 | 0.3081 | 0.1919 | 0.3081 |
|
| 68 |
-
| 0.7147 | 11.2 | 280 | 0.7018 | 0.5006 | 0.1629 | 0.3378 | 0.1622 | 0.3371 |
|
| 69 |
-
| 0.7238 | 12.0 | 300 | 0.7011 | 0.4987 | 0.1768 | 0.3220 | 0.1780 | 0.3232 |
|
| 70 |
-
| 0.7146 | 12.8 | 320 | 0.7013 | 0.4962 | 0.1667 | 0.3295 | 0.1705 | 0.3333 |
|
| 71 |
-
| 0.7108 | 13.6 | 340 | 0.7022 | 0.5019 | 0.1427 | 0.3592 | 0.1408 | 0.3573 |
|
| 72 |
-
| 0.7213 | 14.4 | 360 | 0.6999 | 0.4937 | 0.2090 | 0.2847 | 0.2153 | 0.2910 |
|
| 73 |
-
| 0.7202 | 15.2 | 380 | 0.7011 | 0.4962 | 0.1534 | 0.3428 | 0.1572 | 0.3466 |
|
| 74 |
-
| 0.7136 | 16.0 | 400 | 0.7012 | 0.4981 | 0.1521 | 0.3460 | 0.1540 | 0.3479 |
|
| 75 |
-
| 0.7204 | 16.8 | 420 | 0.7007 | 0.5 | 0.1641 | 0.3359 | 0.1641 | 0.3359 |
|
| 76 |
-
| 0.7238 | 17.6 | 440 | 0.7004 | 0.4962 | 0.1673 | 0.3289 | 0.1711 | 0.3327 |
|
| 77 |
-
| 0.7304 | 18.4 | 460 | 0.7002 | 0.4931 | 0.1705 | 0.3226 | 0.1774 | 0.3295 |
|
| 78 |
-
| 0.7037 | 19.2 | 480 | 0.7003 | 0.4931 | 0.1648 | 0.3283 | 0.1717 | 0.3352 |
|
| 79 |
-
| 0.7228 | 20.0 | 500 | 0.6998 | 0.4956 | 0.1818 | 0.3138 | 0.1862 | 0.3182 |
|
| 80 |
-
| 0.7235 | 20.8 | 520 | 0.6997 | 0.4905 | 0.1761 | 0.3144 | 0.1856 | 0.3239 |
|
| 81 |
-
| 0.7137 | 21.6 | 540 | 0.6993 | 0.4962 | 0.2014 | 0.2948 | 0.2052 | 0.2986 |
|
| 82 |
-
| 0.7152 | 22.4 | 560 | 0.6993 | 0.4931 | 0.1963 | 0.2967 | 0.2033 | 0.3037 |
|
| 83 |
-
| 0.7082 | 23.2 | 580 | 0.6992 | 0.4943 | 0.2014 | 0.2929 | 0.2071 | 0.2986 |
|
| 84 |
-
| 0.7075 | 24.0 | 600 | 0.6995 | 0.4962 | 0.1774 | 0.3188 | 0.1812 | 0.3226 |
|
| 85 |
-
| 0.7145 | 24.8 | 620 | 0.6991 | 0.4924 | 0.1976 | 0.2948 | 0.2052 | 0.3024 |
|
| 86 |
-
| 0.7076 | 25.6 | 640 | 0.7007 | 0.4981 | 0.1332 | 0.3649 | 0.1351 | 0.3668 |
|
| 87 |
-
| 0.7117 | 26.4 | 660 | 0.6998 | 0.4981 | 0.1597 | 0.3384 | 0.1616 | 0.3403 |
|
| 88 |
-
| 0.696 | 27.2 | 680 | 0.6997 | 0.4987 | 0.1591 | 0.3396 | 0.1604 | 0.3409 |
|
| 89 |
-
| 0.7176 | 28.0 | 700 | 0.6994 | 0.4981 | 0.1692 | 0.3289 | 0.1711 | 0.3308 |
|
| 90 |
-
| 0.7148 | 28.8 | 720 | 0.6997 | 0.4981 | 0.1578 | 0.3403 | 0.1597 | 0.3422 |
|
| 91 |
-
| 0.7147 | 29.6 | 740 | 0.6986 | 0.4981 | 0.2159 | 0.2822 | 0.2178 | 0.2841 |
|
| 92 |
-
| 0.7281 | 30.4 | 760 | 0.6989 | 0.4956 | 0.1862 | 0.3093 | 0.1907 | 0.3138 |
|
| 93 |
-
| 0.7052 | 31.2 | 780 | 0.6995 | 0.4981 | 0.1591 | 0.3390 | 0.1610 | 0.3409 |
|
| 94 |
-
| 0.7301 | 32.0 | 800 | 0.6991 | 0.4975 | 0.1749 | 0.3226 | 0.1774 | 0.3251 |
|
| 95 |
-
| 0.7222 | 32.8 | 820 | 0.6988 | 0.4962 | 0.1850 | 0.3112 | 0.1888 | 0.3150 |
|
| 96 |
-
| 0.7202 | 33.6 | 840 | 0.6991 | 0.4949 | 0.1679 | 0.3270 | 0.1730 | 0.3321 |
|
| 97 |
-
| 0.7104 | 34.4 | 860 | 0.6985 | 0.4937 | 0.2014 | 0.2923 | 0.2077 | 0.2986 |
|
| 98 |
-
| 0.7207 | 35.2 | 880 | 0.6982 | 0.4937 | 0.2222 | 0.2715 | 0.2285 | 0.2778 |
|
| 99 |
-
| 0.6985 | 36.0 | 900 | 0.6983 | 0.4981 | 0.2115 | 0.2866 | 0.2134 | 0.2885 |
|
| 100 |
-
| 0.7073 | 36.8 | 920 | 0.6987 | 0.4975 | 0.1806 | 0.3169 | 0.1831 | 0.3194 |
|
| 101 |
-
| 0.7186 | 37.6 | 940 | 0.6983 | 0.4981 | 0.2058 | 0.2923 | 0.2077 | 0.2942 |
|
| 102 |
-
| 0.7084 | 38.4 | 960 | 0.6986 | 0.4956 | 0.1837 | 0.3119 | 0.1881 | 0.3163 |
|
| 103 |
-
| 0.7197 | 39.2 | 980 | 0.6988 | 0.4987 | 0.1742 | 0.3245 | 0.1755 | 0.3258 |
|
| 104 |
-
| 0.7137 | 40.0 | 1000 | 0.6987 | 0.4962 | 0.1761 | 0.3201 | 0.1799 | 0.3239 |
|
| 105 |
-
| 0.7107 | 40.8 | 1020 | 0.6987 | 0.4968 | 0.1730 | 0.3239 | 0.1761 | 0.3270 |
|
| 106 |
-
| 0.6927 | 41.6 | 1040 | 0.6988 | 0.4949 | 0.1692 | 0.3258 | 0.1742 | 0.3308 |
|
| 107 |
-
| 0.7125 | 42.4 | 1060 | 0.6989 | 0.4937 | 0.1610 | 0.3327 | 0.1673 | 0.3390 |
|
| 108 |
-
| 0.7038 | 43.2 | 1080 | 0.6990 | 0.4937 | 0.1572 | 0.3365 | 0.1635 | 0.3428 |
|
| 109 |
-
| 0.7254 | 44.0 | 1100 | 0.6990 | 0.4943 | 0.1578 | 0.3365 | 0.1635 | 0.3422 |
|
| 110 |
-
| 0.7122 | 44.8 | 1120 | 0.6990 | 0.4937 | 0.1559 | 0.3378 | 0.1622 | 0.3441 |
|
| 111 |
-
| 0.7129 | 45.6 | 1140 | 0.6990 | 0.4937 | 0.1553 | 0.3384 | 0.1616 | 0.3447 |
|
| 112 |
-
| 0.7204 | 46.4 | 1160 | 0.6988 | 0.4924 | 0.1635 | 0.3289 | 0.1711 | 0.3365 |
|
| 113 |
-
| 0.716 | 47.2 | 1180 | 0.6987 | 0.4962 | 0.1717 | 0.3245 | 0.1755 | 0.3283 |
|
| 114 |
-
| 0.7192 | 48.0 | 1200 | 0.6987 | 0.4956 | 0.1705 | 0.3251 | 0.1749 | 0.3295 |
|
| 115 |
-
| 0.704 | 48.8 | 1220 | 0.6987 | 0.4956 | 0.1705 | 0.3251 | 0.1749 | 0.3295 |
|
| 116 |
-
| 0.7072 | 49.6 | 1240 | 0.6987 | 0.4924 | 0.1660 | 0.3264 | 0.1736 | 0.3340 |
|
| 117 |
-
|
| 118 |
-
|
| 119 |
### Framework versions
|
| 120 |
|
| 121 |
- Transformers 4.26.1
|
|
|
|
| 2 |
license: mit
|
| 3 |
tags:
|
| 4 |
- generated_from_trainer
|
|
|
|
|
|
|
| 5 |
model-index:
|
| 6 |
- name: gpt2_winobias_classifieronly
|
| 7 |
results: []
|
|
|
|
| 12 |
|
| 13 |
# gpt2_winobias_classifieronly
|
| 14 |
|
| 15 |
+
This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16 |
|
| 17 |
## Model description
|
| 18 |
|
|
|
|
| 39 |
- lr_scheduler_type: linear
|
| 40 |
- num_epochs: 50
|
| 41 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
### Framework versions
|
| 43 |
|
| 44 |
- Transformers 4.26.1
|