Training in progress, epoch 1
Browse files- all_results.json +14 -0
- eval_results.json +10 -0
- predict_results.json +9 -0
- test_predictions.csv +100 -0
- training_args.bin +1 -1
all_results.json
CHANGED
|
@@ -1,5 +1,19 @@
|
|
| 1 |
{
|
| 2 |
"epoch": 4.0,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
"total_flos": 2512343535714304.0,
|
| 4 |
"train_loss": 0.3421715199947357,
|
| 5 |
"train_runtime": 21.295,
|
|
|
|
| 1 |
{
|
| 2 |
"epoch": 4.0,
|
| 3 |
+
"eval_loss": 0.41252821683883667,
|
| 4 |
+
"eval_macro-f1": 0.029029793735676088,
|
| 5 |
+
"eval_micro-f1": 0.16521739130434782,
|
| 6 |
+
"eval_runtime": 1.226,
|
| 7 |
+
"eval_samples": 100,
|
| 8 |
+
"eval_samples_per_second": 81.568,
|
| 9 |
+
"eval_steps_per_second": 3.263,
|
| 10 |
+
"predict_loss": 0.39896854758262634,
|
| 11 |
+
"predict_macro-f1": 0.01948051948051948,
|
| 12 |
+
"predict_micro-f1": 0.1095890410958904,
|
| 13 |
+
"predict_runtime": 0.9572,
|
| 14 |
+
"predict_samples": 100,
|
| 15 |
+
"predict_samples_per_second": 104.467,
|
| 16 |
+
"predict_steps_per_second": 4.179,
|
| 17 |
"total_flos": 2512343535714304.0,
|
| 18 |
"train_loss": 0.3421715199947357,
|
| 19 |
"train_runtime": 21.295,
|
eval_results.json
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 4.0,
|
| 3 |
+
"eval_loss": 0.41252821683883667,
|
| 4 |
+
"eval_macro-f1": 0.029029793735676088,
|
| 5 |
+
"eval_micro-f1": 0.16521739130434782,
|
| 6 |
+
"eval_runtime": 1.226,
|
| 7 |
+
"eval_samples": 100,
|
| 8 |
+
"eval_samples_per_second": 81.568,
|
| 9 |
+
"eval_steps_per_second": 3.263
|
| 10 |
+
}
|
predict_results.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"predict_loss": 0.39896854758262634,
|
| 3 |
+
"predict_macro-f1": 0.01948051948051948,
|
| 4 |
+
"predict_micro-f1": 0.1095890410958904,
|
| 5 |
+
"predict_runtime": 0.9572,
|
| 6 |
+
"predict_samples": 100,
|
| 7 |
+
"predict_samples_per_second": 104.467,
|
| 8 |
+
"predict_steps_per_second": 4.179
|
| 9 |
+
}
|
test_predictions.csv
ADDED
|
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
0 -0.95752 -1.46875 -1.03516 -0.21106 -1.07812 -1.48047 -0.77197 -0.96191 -1.46191 -1.30762
|
| 2 |
+
1 -1.04492 -1.48730 -1.06738 -0.14148 -1.15820 -1.57324 -0.73584 -0.95996 -1.56445 -1.25879
|
| 3 |
+
2 -1.00586 -1.48535 -1.03516 -0.20349 -1.17090 -1.53906 -0.70801 -0.96777 -1.62891 -1.16016
|
| 4 |
+
3 -1.03027 -1.47070 -1.04297 -0.18628 -1.18652 -1.53613 -0.72266 -0.94824 -1.63477 -1.16602
|
| 5 |
+
4 -1.05762 -1.48047 -1.05566 -0.17114 -1.15625 -1.54297 -0.73682 -0.94971 -1.58887 -1.24805
|
| 6 |
+
5 -1.02148 -1.48242 -1.04590 -0.20508 -1.10254 -1.51270 -0.77148 -0.95459 -1.49902 -1.28223
|
| 7 |
+
6 -1.05762 -1.47461 -1.05957 -0.15637 -1.17383 -1.55176 -0.73828 -0.95215 -1.59180 -1.25391
|
| 8 |
+
7 -1.04102 -1.48047 -1.04980 -0.19666 -1.11035 -1.52441 -0.76807 -0.95264 -1.51172 -1.25879
|
| 9 |
+
8 -1.06348 -1.49707 -1.04297 -0.18201 -1.16309 -1.53711 -0.72998 -0.95996 -1.60645 -1.24316
|
| 10 |
+
9 -1.06836 -1.48730 -1.08105 -0.17944 -1.17969 -1.53125 -0.70996 -0.95752 -1.62207 -1.17285
|
| 11 |
+
10 -1.01367 -1.47168 -1.04785 -0.18591 -1.17578 -1.55371 -0.71338 -0.95215 -1.64648 -1.16504
|
| 12 |
+
11 -1.05371 -1.48242 -1.07617 -0.18286 -1.17578 -1.53809 -0.70996 -0.95898 -1.63770 -1.17773
|
| 13 |
+
12 -1.02148 -1.47559 -1.02637 -0.21130 -1.16797 -1.51367 -0.67529 -0.94971 -1.61621 -1.12891
|
| 14 |
+
13 -1.06348 -1.48242 -1.04395 -0.16455 -1.16211 -1.54004 -0.72168 -0.96094 -1.58887 -1.24805
|
| 15 |
+
14 -0.94629 -1.44531 -1.00391 -0.22925 -1.03418 -1.45703 -0.76611 -0.91553 -1.43652 -1.30859
|
| 16 |
+
15 -1.06250 -1.49609 -1.04492 -0.17432 -1.16406 -1.53027 -0.73730 -0.95557 -1.59277 -1.24316
|
| 17 |
+
16 -1.01465 -1.47949 -1.03125 -0.18591 -1.16113 -1.54199 -0.67676 -0.95947 -1.61523 -1.13965
|
| 18 |
+
17 -1.05469 -1.47559 -1.06836 -0.18933 -1.16406 -1.55371 -0.71338 -0.96680 -1.61523 -1.19238
|
| 19 |
+
18 -1.07227 -1.49512 -1.04297 -0.19360 -1.17480 -1.53711 -0.73438 -0.95410 -1.60352 -1.23535
|
| 20 |
+
19 -1.03125 -1.48145 -1.06543 -0.19080 -1.17383 -1.55176 -0.70605 -0.94629 -1.65137 -1.16602
|
| 21 |
+
20 -0.94043 -1.47168 -0.99219 -0.21948 -1.02344 -1.43848 -0.77051 -0.92041 -1.41895 -1.30078
|
| 22 |
+
21 -0.86475 -1.38574 -0.93018 -0.27197 -0.83887 -1.37988 -0.70605 -0.73389 -1.30078 -1.16602
|
| 23 |
+
22 -1.05176 -1.47656 -1.04980 -0.16187 -1.11523 -1.54688 -0.75977 -0.97217 -1.53809 -1.27148
|
| 24 |
+
23 -1.03125 -1.48145 -1.07031 -0.18848 -1.19141 -1.54004 -0.72266 -0.95020 -1.62891 -1.17676
|
| 25 |
+
24 -1.04102 -1.49707 -1.07910 -0.18237 -1.18457 -1.53906 -0.71582 -0.94385 -1.62793 -1.17480
|
| 26 |
+
25 -1.04688 -1.47656 -1.05078 -0.19092 -1.10156 -1.53613 -0.77344 -0.96875 -1.52637 -1.27344
|
| 27 |
+
26 -1.05762 -1.48047 -1.05664 -0.15576 -1.14355 -1.56738 -0.72803 -0.95850 -1.55664 -1.26172
|
| 28 |
+
27 -1.07129 -1.49023 -1.07129 -0.19519 -1.14941 -1.55078 -0.72559 -0.96240 -1.60352 -1.20410
|
| 29 |
+
28 -1.01758 -1.47461 -1.04785 -0.19104 -1.19434 -1.54785 -0.71582 -0.94775 -1.63379 -1.17773
|
| 30 |
+
29 -1.06250 -1.47559 -1.08203 -0.18909 -1.16602 -1.52930 -0.70996 -0.96631 -1.63086 -1.19043
|
| 31 |
+
30 -1.00879 -1.49023 -1.04688 -0.20374 -1.16992 -1.56055 -0.70215 -0.95557 -1.62988 -1.14844
|
| 32 |
+
31 -0.96729 -1.46094 -1.02930 -0.20190 -1.09277 -1.50098 -0.76660 -0.95605 -1.47559 -1.30078
|
| 33 |
+
32 -1.04883 -1.49414 -1.06738 -0.17944 -1.18457 -1.52734 -0.71045 -0.94141 -1.63770 -1.16797
|
| 34 |
+
33 -1.01758 -1.48926 -1.04492 -0.19751 -1.17871 -1.54688 -0.70264 -0.96973 -1.63867 -1.15723
|
| 35 |
+
34 -1.01465 -1.46289 -1.01855 -0.22412 -1.17090 -1.50195 -0.65820 -0.94482 -1.61621 -1.10059
|
| 36 |
+
35 -1.06543 -1.48535 -1.07520 -0.19128 -1.16113 -1.55664 -0.73145 -0.95752 -1.61426 -1.20215
|
| 37 |
+
36 -1.05078 -1.48828 -1.06641 -0.17871 -1.17871 -1.53125 -0.70703 -0.93701 -1.62598 -1.16309
|
| 38 |
+
37 -1.00195 -1.48730 -1.03027 -0.20764 -1.16699 -1.54199 -0.70166 -0.95557 -1.63770 -1.14746
|
| 39 |
+
38 -1.05469 -1.48047 -1.06738 -0.18994 -1.16406 -1.53809 -0.70361 -0.96240 -1.62305 -1.18555
|
| 40 |
+
39 -0.98438 -1.48438 -1.04785 -0.20654 -1.09082 -1.48340 -0.75684 -0.95215 -1.48047 -1.29199
|
| 41 |
+
40 -1.06250 -1.48340 -1.07812 -0.18811 -1.17969 -1.53809 -0.70703 -0.95410 -1.62402 -1.16113
|
| 42 |
+
41 -1.06641 -1.48340 -1.04199 -0.18945 -1.15625 -1.53125 -0.73975 -0.95508 -1.59570 -1.24707
|
| 43 |
+
42 -0.91797 -1.46191 -0.95264 -0.24731 -0.97949 -1.41406 -0.77393 -0.91553 -1.38281 -1.33008
|
| 44 |
+
43 -1.05664 -1.47852 -1.06152 -0.18835 -1.14844 -1.55078 -0.72803 -0.95068 -1.61230 -1.20508
|
| 45 |
+
44 -1.04785 -1.47070 -1.05469 -0.15747 -1.15430 -1.55273 -0.72998 -0.96289 -1.58984 -1.23828
|
| 46 |
+
45 -1.05762 -1.48730 -1.05371 -0.18872 -1.15430 -1.54395 -0.73828 -0.96045 -1.58887 -1.22656
|
| 47 |
+
46 -1.07031 -1.47949 -1.04590 -0.18335 -1.16504 -1.53906 -0.74658 -0.95410 -1.60059 -1.24414
|
| 48 |
+
47 -1.01172 -1.48340 -1.04492 -0.20618 -1.10156 -1.49512 -0.76465 -0.97168 -1.50293 -1.28613
|
| 49 |
+
48 -1.06055 -1.47559 -1.07617 -0.18481 -1.15332 -1.56445 -0.71094 -0.97168 -1.60840 -1.20703
|
| 50 |
+
49 -1.03906 -1.47949 -1.06152 -0.15381 -1.15430 -1.56934 -0.74902 -0.96826 -1.54980 -1.26465
|
| 51 |
+
50 -1.00488 -1.47266 -1.04590 -0.18469 -1.17188 -1.55273 -0.71436 -0.94824 -1.64648 -1.17578
|
| 52 |
+
51 -1.02441 -1.47656 -1.02637 -0.21460 -1.16309 -1.52051 -0.67383 -0.95312 -1.62207 -1.12695
|
| 53 |
+
52 -0.97803 -1.47656 -1.03320 -0.21094 -1.10352 -1.48633 -0.75830 -0.95117 -1.49805 -1.29199
|
| 54 |
+
53 -0.88574 -1.38574 -0.93652 -0.26001 -0.84912 -1.37012 -0.72168 -0.72656 -1.28418 -1.16211
|
| 55 |
+
54 -1.02051 -1.48242 -1.02734 -0.21558 -1.16992 -1.52148 -0.67139 -0.95020 -1.61914 -1.12402
|
| 56 |
+
55 -0.94678 -1.46289 -1.02930 -0.21960 -1.05566 -1.48535 -0.78613 -0.95166 -1.46289 -1.30859
|
| 57 |
+
56 -1.02344 -1.48438 -1.02930 -0.21472 -1.16504 -1.52344 -0.66797 -0.94434 -1.61328 -1.11719
|
| 58 |
+
57 -1.04004 -1.49414 -1.06250 -0.16211 -1.13672 -1.55566 -0.75342 -0.95020 -1.54688 -1.27148
|
| 59 |
+
58 -1.03320 -1.48242 -1.05469 -0.20056 -1.11523 -1.53027 -0.77441 -0.95166 -1.51465 -1.28027
|
| 60 |
+
59 -1.02051 -1.47363 -1.03809 -0.20361 -1.17090 -1.52539 -0.69531 -0.95117 -1.62109 -1.13379
|
| 61 |
+
60 -0.92871 -1.46777 -0.94580 -0.25464 -0.98145 -1.41602 -0.76758 -0.90674 -1.38184 -1.33594
|
| 62 |
+
61 -1.03418 -1.48340 -1.05371 -0.18127 -1.10840 -1.54492 -0.77783 -0.95752 -1.51172 -1.27734
|
| 63 |
+
62 -1.00684 -1.48535 -1.03125 -0.20093 -1.09570 -1.49316 -0.76221 -0.97656 -1.49902 -1.28027
|
| 64 |
+
63 -1.05176 -1.46191 -1.05957 -0.14258 -1.16016 -1.56250 -0.73975 -0.95605 -1.56836 -1.25391
|
| 65 |
+
64 -1.05273 -1.48633 -1.06250 -0.19092 -1.15332 -1.54980 -0.74463 -0.95020 -1.59570 -1.21875
|
| 66 |
+
65 -1.04199 -1.49023 -1.06250 -0.17859 -1.18164 -1.52539 -0.70264 -0.95215 -1.62988 -1.17578
|
| 67 |
+
66 -1.03711 -1.47168 -1.06055 -0.19019 -1.10840 -1.54688 -0.77002 -0.96240 -1.51660 -1.27539
|
| 68 |
+
67 -1.05762 -1.49023 -1.04883 -0.17651 -1.15137 -1.53418 -0.72314 -0.95654 -1.60645 -1.24121
|
| 69 |
+
68 -1.05273 -1.47070 -1.05566 -0.15625 -1.14746 -1.56445 -0.74121 -0.96436 -1.54785 -1.26562
|
| 70 |
+
69 -1.01562 -1.49707 -1.04590 -0.19934 -1.08398 -1.49512 -0.77783 -0.96289 -1.50000 -1.28320
|
| 71 |
+
70 -1.00977 -1.48438 -1.04688 -0.19067 -1.09570 -1.51270 -0.76709 -0.95752 -1.51367 -1.28613
|
| 72 |
+
71 -1.02637 -1.47949 -1.02051 -0.21326 -1.16309 -1.52051 -0.67139 -0.96045 -1.62402 -1.11914
|
| 73 |
+
72 -1.03320 -1.48438 -1.06641 -0.18115 -1.19531 -1.54492 -0.71729 -0.94531 -1.63574 -1.17871
|
| 74 |
+
73 -1.02051 -1.47656 -1.03125 -0.20496 -1.17285 -1.53320 -0.69727 -0.94043 -1.62500 -1.13965
|
| 75 |
+
74 -1.06836 -1.46484 -1.06836 -0.18311 -1.15918 -1.54102 -0.70117 -0.96875 -1.61621 -1.17969
|
| 76 |
+
75 -1.01465 -1.48145 -1.04395 -0.20996 -1.17285 -1.54785 -0.68555 -0.95410 -1.61914 -1.13184
|
| 77 |
+
76 -1.04980 -1.47168 -1.08105 -0.18384 -1.17090 -1.54004 -0.70703 -0.95312 -1.62988 -1.17480
|
| 78 |
+
77 -1.04883 -1.48633 -1.05859 -0.16431 -1.12695 -1.56348 -0.75684 -0.96680 -1.53223 -1.27344
|
| 79 |
+
78 -1.01758 -1.48828 -1.03906 -0.19238 -1.18652 -1.55859 -0.71191 -0.95312 -1.64355 -1.16699
|
| 80 |
+
79 -1.01562 -1.47168 -1.02637 -0.20325 -1.16504 -1.54004 -0.69434 -0.96094 -1.62012 -1.13965
|
| 81 |
+
80 -1.06250 -1.49316 -1.05371 -0.20068 -1.17578 -1.54004 -0.73389 -0.95752 -1.59668 -1.21973
|
| 82 |
+
81 -0.97021 -1.47461 -1.03223 -0.21094 -1.10059 -1.50000 -0.76123 -0.94971 -1.48438 -1.30859
|
| 83 |
+
82 -1.02637 -1.46680 -1.02734 -0.22961 -1.17285 -1.51465 -0.65576 -0.94678 -1.62305 -1.08301
|
| 84 |
+
83 -0.88037 -1.38086 -0.94385 -0.27905 -0.86670 -1.38184 -0.71826 -0.73193 -1.30176 -1.18164
|
| 85 |
+
84 -1.06055 -1.48438 -1.05176 -0.17139 -1.16016 -1.55176 -0.73730 -0.95117 -1.58789 -1.24121
|
| 86 |
+
85 -1.06738 -1.49512 -1.03809 -0.17310 -1.16504 -1.53809 -0.74463 -0.95166 -1.59375 -1.23828
|
| 87 |
+
86 -1.04883 -1.47656 -1.06934 -0.15308 -1.12793 -1.55176 -0.75146 -0.96338 -1.53418 -1.27539
|
| 88 |
+
87 -1.07422 -1.50195 -1.05469 -0.18311 -1.17871 -1.53809 -0.75244 -0.95166 -1.58887 -1.23828
|
| 89 |
+
88 -1.03809 -1.47168 -1.04785 -0.17590 -1.12012 -1.54785 -0.77490 -0.96338 -1.53516 -1.26074
|
| 90 |
+
89 -1.01172 -1.47754 -1.04297 -0.20667 -1.15918 -1.54590 -0.70215 -0.96680 -1.63770 -1.13965
|
| 91 |
+
90 -1.02246 -1.47363 -1.01465 -0.23584 -1.16602 -1.50195 -0.64893 -0.93213 -1.62207 -1.08691
|
| 92 |
+
91 -1.05371 -1.46875 -1.07129 -0.18909 -1.18262 -1.54102 -0.70703 -0.95752 -1.62598 -1.16797
|
| 93 |
+
92 -1.01367 -1.47949 -1.04297 -0.19666 -1.15820 -1.52539 -0.69482 -0.95557 -1.62695 -1.12695
|
| 94 |
+
93 -1.06152 -1.47559 -1.05859 -0.15112 -1.12207 -1.54688 -0.75146 -0.96533 -1.55176 -1.28125
|
| 95 |
+
94 -1.05762 -1.49219 -1.05566 -0.14966 -1.14551 -1.55176 -0.75146 -0.96240 -1.55762 -1.25879
|
| 96 |
+
95 -0.90674 -1.46191 -0.91357 -0.26904 -0.94580 -1.40137 -0.79736 -0.87793 -1.34863 -1.31934
|
| 97 |
+
96 -0.87451 -1.38574 -0.93799 -0.28394 -0.84326 -1.37500 -0.72070 -0.74805 -1.28125 -1.16699
|
| 98 |
+
97 -0.97461 -1.47656 -1.03027 -0.21521 -1.07520 -1.48047 -0.77148 -0.95557 -1.47754 -1.29297
|
| 99 |
+
98 -1.03516 -1.48047 -1.03711 -0.18958 -1.11035 -1.53418 -0.76172 -0.95557 -1.51660 -1.27832
|
| 100 |
+
99 -0.93164 -1.46094 -1.00586 -0.23853 -1.02051 -1.43945 -0.77100 -0.92871 -1.41992 -1.29688
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5368
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1eb831cb1a1fa30eb821d46dc3e3b4fd594ab220bb277cf54ee0abd97b381bf4
|
| 3 |
size 5368
|