trainer: training complete at 2024-10-26 20:58:24.147472.
Browse files- README.md +23 -37
- meta_data/README_s42_e5.md +26 -24
- meta_data/meta_s42_e5_cvi0.json +1 -1
- model.safetensors +1 -1
README.md
CHANGED
|
@@ -1,10 +1,11 @@
|
|
| 1 |
---
|
|
|
|
| 2 |
license: apache-2.0
|
| 3 |
base_model: allenai/longformer-base-4096
|
| 4 |
tags:
|
| 5 |
- generated_from_trainer
|
| 6 |
datasets:
|
| 7 |
-
-
|
| 8 |
metrics:
|
| 9 |
- accuracy
|
| 10 |
model-index:
|
|
@@ -14,15 +15,15 @@ model-index:
|
|
| 14 |
name: Token Classification
|
| 15 |
type: token-classification
|
| 16 |
dataset:
|
| 17 |
-
name:
|
| 18 |
-
type:
|
| 19 |
config: spans
|
| 20 |
split: train[0%:20%]
|
| 21 |
args: spans
|
| 22 |
metrics:
|
| 23 |
- name: Accuracy
|
| 24 |
type: accuracy
|
| 25 |
-
value: 0.
|
| 26 |
---
|
| 27 |
|
| 28 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
@@ -30,15 +31,15 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 30 |
|
| 31 |
# longformer-spans
|
| 32 |
|
| 33 |
-
This model is a fine-tuned version of [allenai/longformer-base-4096](https://huggingface.co/allenai/longformer-base-4096) on the
|
| 34 |
It achieves the following results on the evaluation set:
|
| 35 |
-
- Loss: 0.
|
| 36 |
-
- B: {'precision': 0.
|
| 37 |
-
- I: {'precision': 0.
|
| 38 |
-
- O: {'precision': 0.
|
| 39 |
-
- Accuracy: 0.
|
| 40 |
-
- Macro avg: {'precision': 0.
|
| 41 |
-
- Weighted avg: {'precision': 0.
|
| 42 |
|
| 43 |
## Model description
|
| 44 |
|
|
@@ -63,37 +64,22 @@ The following hyperparameters were used during training:
|
|
| 63 |
- seed: 42
|
| 64 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 65 |
- lr_scheduler_type: linear
|
| 66 |
-
- num_epochs:
|
| 67 |
|
| 68 |
### Training results
|
| 69 |
|
| 70 |
| Training Loss | Epoch | Step | Validation Loss | B | I | O | Accuracy | Macro avg | Weighted avg |
|
| 71 |
|:-------------:|:-----:|:----:|:---------------:|:------------------------------------------------------------------------------------------------------------------:|:-------------------------------------------------------------------------------------------------------------------:|:------------------------------------------------------------------------------------------------------------------:|:--------:|:-------------------------------------------------------------------------------------------------------------------:|:-------------------------------------------------------------------------------------------------------------------:|
|
| 72 |
-
| No log | 1.0 |
|
| 73 |
-
| No log | 2.0 |
|
| 74 |
-
| No log | 3.0 |
|
| 75 |
-
| No log | 4.0 |
|
| 76 |
-
| No log | 5.0 |
|
| 77 |
-
| No log | 6.0 | 486 | 0.2930 | {'precision': 0.841927303465765, 'recall': 0.8790820829655781, 'f1-score': 0.8601036269430052, 'support': 1133.0} | {'precision': 0.9452679589509693, 'recall': 0.9495990836197021, 'f1-score': 0.9474285714285714, 'support': 18333.0} | {'precision': 0.9045613314156564, 'recall': 0.8922780705310093, 'f1-score': 0.8983777165595348, 'support': 9868.0} | 0.9276 | {'precision': 0.8972521979441302, 'recall': 0.9069864123720964, 'f1-score': 0.9019699716437038, 'support': 29334.0} | {'precision': 0.9275827485063247, 'recall': 0.9275925547146656, 'f1-score': 0.9275549436263691, 'support': 29334.0} |
|
| 78 |
-
| 0.1621 | 7.0 | 567 | 0.3149 | {'precision': 0.8406639004149378, 'recall': 0.8940864960282436, 'f1-score': 0.8665526090675792, 'support': 1133.0} | {'precision': 0.9382959450098577, 'recall': 0.9605083728795069, 'f1-score': 0.9492722371967655, 'support': 18333.0} | {'precision': 0.9227729117709891, 'recall': 0.8754560194568302, 'f1-score': 0.8984919396775871, 'support': 9868.0} | 0.9293 | {'precision': 0.9005775857319281, 'recall': 0.9100169627881934, 'f1-score': 0.9047722619806439, 'support': 29334.0} | {'precision': 0.9293030221719495, 'recall': 0.9293311515647371, 'f1-score': 0.9289946986889036, 'support': 29334.0} |
|
| 79 |
-
| 0.1621 | 8.0 | 648 | 0.3477 | {'precision': 0.8284552845528456, 'recall': 0.8993821712268314, 'f1-score': 0.8624629707998307, 'support': 1133.0} | {'precision': 0.9356556940449557, 'recall': 0.9581628756886489, 'f1-score': 0.9467755410030451, 'support': 18333.0} | {'precision': 0.9191854233654877, 'recall': 0.8690717470612079, 'f1-score': 0.8934263985831857, 'support': 9868.0} | 0.9259 | {'precision': 0.8944321339877629, 'recall': 0.9088722646588961, 'f1-score': 0.9008883034620205, 'support': 29334.0} | {'precision': 0.9259745494680297, 'recall': 0.9259221381332242, 'f1-score': 0.9255723133682386, 'support': 29334.0} |
|
| 80 |
-
| 0.1621 | 9.0 | 729 | 0.3808 | {'precision': 0.8464135021097047, 'recall': 0.8852603706972639, 'f1-score': 0.8654012079378774, 'support': 1133.0} | {'precision': 0.9316216786166175, 'recall': 0.9638902525500463, 'f1-score': 0.9474813007694164, 'support': 18333.0} | {'precision': 0.9268053588933667, 'recall': 0.8622821240372922, 'f1-score': 0.8933802299333298, 'support': 9868.0} | 0.9267 | {'precision': 0.9016135132065629, 'recall': 0.9038109157615342, 'f1-score': 0.9020875795468745, 'support': 29334.0} | {'precision': 0.9267103706800465, 'recall': 0.9266721210881571, 'f1-score': 0.9261113508073029, 'support': 29334.0} |
|
| 81 |
-
| 0.1621 | 10.0 | 810 | 0.4663 | {'precision': 0.8380872483221476, 'recall': 0.881729920564872, 'f1-score': 0.8593548387096774, 'support': 1133.0} | {'precision': 0.9158687080751703, 'recall': 0.9756177385043364, 'f1-score': 0.9447995351539802, 'support': 18333.0} | {'precision': 0.9469406710786021, 'recall': 0.8265099310903932, 'f1-score': 0.8826362209837131, 'support': 9868.0} | 0.9218 | {'precision': 0.9002988758253068, 'recall': 0.8946191967198672, 'f1-score': 0.8955968649491236, 'support': 29334.0} | {'precision': 0.9233171207368492, 'recall': 0.9218313220154087, 'f1-score': 0.9205874800198834, 'support': 29334.0} |
|
| 82 |
-
| 0.1621 | 11.0 | 891 | 0.3998 | {'precision': 0.8421052631578947, 'recall': 0.8755516328331863, 'f1-score': 0.8585028126352229, 'support': 1133.0} | {'precision': 0.941814648890808, 'recall': 0.9517809414716631, 'f1-score': 0.9467715680954965, 'support': 18333.0} | {'precision': 0.906636203136359, 'recall': 0.8846777462505067, 'f1-score': 0.8955223880597014, 'support': 9868.0} | 0.9263 | {'precision': 0.8968520383950206, 'recall': 0.9040034401851186, 'f1-score': 0.9002655895968069, 'support': 29334.0} | {'precision': 0.9261293813943774, 'recall': 0.9262630394763756, 'f1-score': 0.9261219666592889, 'support': 29334.0} |
|
| 83 |
-
| 0.1621 | 12.0 | 972 | 0.4524 | {'precision': 0.8503401360544217, 'recall': 0.8826125330979699, 'f1-score': 0.8661758336942399, 'support': 1133.0} | {'precision': 0.9383342231713828, 'recall': 0.9586537937053401, 'f1-score': 0.9483851819874267, 'support': 18333.0} | {'precision': 0.9182223165040305, 'recall': 0.8772800972841508, 'f1-score': 0.8972844112769486, 'support': 9868.0} | 0.9283 | {'precision': 0.9022988919099451, 'recall': 0.906182141362487, 'f1-score': 0.9039484756528716, 'support': 29334.0} | {'precision': 0.9281698543264606, 'recall': 0.9283425376695984, 'f1-score': 0.9280195449455239, 'support': 29334.0} |
|
| 84 |
-
| 0.0212 | 13.0 | 1053 | 0.4537 | {'precision': 0.8431703204047217, 'recall': 0.8826125330979699, 'f1-score': 0.8624407072013798, 'support': 1133.0} | {'precision': 0.9365968111768783, 'recall': 0.9580537827960508, 'f1-score': 0.94720379658092, 'support': 18333.0} | {'precision': 0.9167642362959021, 'recall': 0.8728212403729225, 'f1-score': 0.8942532315838654, 'support': 9868.0} | 0.9265 | {'precision': 0.8988437892925006, 'recall': 0.9044958520889811, 'f1-score': 0.9012992451220551, 'support': 29334.0} | {'precision': 0.926316588126141, 'recall': 0.9264675802822663, 'f1-score': 0.9261172500595471, 'support': 29334.0} |
|
| 85 |
-
| 0.0212 | 14.0 | 1134 | 0.4902 | {'precision': 0.8573883161512027, 'recall': 0.880847308031774, 'f1-score': 0.8689595124074879, 'support': 1133.0} | {'precision': 0.9300970873786408, 'recall': 0.9667266677575956, 'f1-score': 0.9480582004921365, 'support': 18333.0} | {'precision': 0.9303346132748217, 'recall': 0.8593433319821646, 'f1-score': 0.8934309645472265, 'support': 9868.0} | 0.9273 | {'precision': 0.9059400056015551, 'recall': 0.9023057692571781, 'f1-score': 0.9034828924822836, 'support': 29334.0} | {'precision': 0.9273686789700647, 'recall': 0.9272857435058294, 'f1-score': 0.9266264019680934, 'support': 29334.0} |
|
| 86 |
-
| 0.0212 | 15.0 | 1215 | 0.4631 | {'precision': 0.8514090520922288, 'recall': 0.8799646954986761, 'f1-score': 0.865451388888889, 'support': 1133.0} | {'precision': 0.943136407819419, 'recall': 0.9526536846124475, 'f1-score': 0.9478711568207105, 'support': 18333.0} | {'precision': 0.9084499740798341, 'recall': 0.8879205512768544, 'f1-score': 0.8980679546968688, 'support': 9868.0} | 0.9281 | {'precision': 0.9009984779971606, 'recall': 0.9068463104626593, 'f1-score': 0.9037968334688228, 'support': 29334.0} | {'precision': 0.9279249527781314, 'recall': 0.9280698165950774, 'f1-score': 0.9279338964530544, 'support': 29334.0} |
|
| 87 |
-
| 0.0212 | 16.0 | 1296 | 0.4685 | {'precision': 0.8621291448516579, 'recall': 0.8720211827007943, 'f1-score': 0.8670469504168494, 'support': 1133.0} | {'precision': 0.9403208556149732, 'recall': 0.9591447117220313, 'f1-score': 0.949639510706667, 'support': 18333.0} | {'precision': 0.917685497470489, 'recall': 0.8823469801378192, 'f1-score': 0.8996693531721429, 'support': 9868.0} | 0.9299 | {'precision': 0.9067118326457067, 'recall': 0.9045042915202149, 'f1-score': 0.9054519380985532, 'support': 29334.0} | {'precision': 0.9296862022276206, 'recall': 0.9299447739824095, 'f1-score': 0.9296394123443894, 'support': 29334.0} |
|
| 88 |
-
| 0.0212 | 17.0 | 1377 | 0.5305 | {'precision': 0.8462823725981621, 'recall': 0.8940864960282436, 'f1-score': 0.8695278969957082, 'support': 1133.0} | {'precision': 0.9287246847035429, 'recall': 0.9680357824687722, 'f1-score': 0.9479728646973986, 'support': 18333.0} | {'precision': 0.9344262295081968, 'recall': 0.8548844750709363, 'f1-score': 0.892887383573243, 'support': 9868.0} | 0.9271 | {'precision': 0.9031444289366339, 'recall': 0.9056689178559841, 'f1-score': 0.9034627150887832, 'support': 29334.0} | {'precision': 0.927458430681484, 'recall': 0.9271152928342538, 'f1-score': 0.9264121612086422, 'support': 29334.0} |
|
| 89 |
-
| 0.0212 | 18.0 | 1458 | 0.5198 | {'precision': 0.847972972972973, 'recall': 0.8861429832303619, 'f1-score': 0.8666378938282262, 'support': 1133.0} | {'precision': 0.9337531086300862, 'recall': 0.9625811378388698, 'f1-score': 0.9479480017189514, 'support': 18333.0} | {'precision': 0.9244406010161064, 'recall': 0.866639643291447, 'f1-score': 0.8946074585490874, 'support': 9868.0} | 0.9274 | {'precision': 0.9020555608730553, 'recall': 0.9051212547868929, 'f1-score': 0.9030644513654217, 'support': 29334.0} | {'precision': 0.9273071851680879, 'recall': 0.9273539237744597, 'f1-score': 0.9268636343554685, 'support': 29334.0} |
|
| 90 |
-
| 0.0055 | 19.0 | 1539 | 0.5277 | {'precision': 0.8447986577181208, 'recall': 0.8887908208296558, 'f1-score': 0.8662365591397849, 'support': 1133.0} | {'precision': 0.9328933474128828, 'recall': 0.9637811596574484, 'f1-score': 0.9480857457140558, 'support': 18333.0} | {'precision': 0.9267550532492936, 'recall': 0.8642075395216863, 'f1-score': 0.8943890928159414, 'support': 9868.0} | 0.9274 | {'precision': 0.9014823527934324, 'recall': 0.9055931733362635, 'f1-score': 0.9029037992232607, 'support': 29334.0} | {'precision': 0.9274258363257326, 'recall': 0.9273880139087748, 'f1-score': 0.9268607610823233, 'support': 29334.0} |
|
| 91 |
-
| 0.0055 | 20.0 | 1620 | 0.5306 | {'precision': 0.8433835845896147, 'recall': 0.8887908208296558, 'f1-score': 0.8654920498495917, 'support': 1133.0} | {'precision': 0.9324545214869496, 'recall': 0.9645993563519337, 'f1-score': 0.9482545981017748, 'support': 18333.0} | {'precision': 0.9282833787465941, 'recall': 0.8630928252938792, 'f1-score': 0.8945019167148034, 'support': 9868.0} | 0.9275 | {'precision': 0.9013738282743861, 'recall': 0.9054943341584897, 'f1-score': 0.90274952155539, 'support': 29334.0} | {'precision': 0.9276110562907095, 'recall': 0.9275243744460353, 'f1-score': 0.9269754876123645, 'support': 29334.0} |
|
| 92 |
|
| 93 |
|
| 94 |
### Framework versions
|
| 95 |
|
| 96 |
-
- Transformers 4.
|
| 97 |
-
- Pytorch 2.
|
| 98 |
-
- Datasets 2.
|
| 99 |
-
- Tokenizers 0.
|
|
|
|
| 1 |
---
|
| 2 |
+
library_name: transformers
|
| 3 |
license: apache-2.0
|
| 4 |
base_model: allenai/longformer-base-4096
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
| 7 |
datasets:
|
| 8 |
+
- stab-gurevych-essays
|
| 9 |
metrics:
|
| 10 |
- accuracy
|
| 11 |
model-index:
|
|
|
|
| 15 |
name: Token Classification
|
| 16 |
type: token-classification
|
| 17 |
dataset:
|
| 18 |
+
name: stab-gurevych-essays
|
| 19 |
+
type: stab-gurevych-essays
|
| 20 |
config: spans
|
| 21 |
split: train[0%:20%]
|
| 22 |
args: spans
|
| 23 |
metrics:
|
| 24 |
- name: Accuracy
|
| 25 |
type: accuracy
|
| 26 |
+
value: 0.9739926865110556
|
| 27 |
---
|
| 28 |
|
| 29 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
|
| 31 |
|
| 32 |
# longformer-spans
|
| 33 |
|
| 34 |
+
This model is a fine-tuned version of [allenai/longformer-base-4096](https://huggingface.co/allenai/longformer-base-4096) on the stab-gurevych-essays dataset.
|
| 35 |
It achieves the following results on the evaluation set:
|
| 36 |
+
- Loss: 0.0856
|
| 37 |
+
- B: {'precision': 0.8861301369863014, 'recall': 0.913503971756399, 'f1-score': 0.8996088657105606, 'support': 1133.0}
|
| 38 |
+
- I: {'precision': 0.9856182499448976, 'recall': 0.978661705969251, 'f1-score': 0.9821276595744681, 'support': 18277.0}
|
| 39 |
+
- O: {'precision': 0.963097033685269, 'recall': 0.9722870774540656, 'f1-score': 0.9676702364113963, 'support': 9851.0}
|
| 40 |
+
- Accuracy: 0.9740
|
| 41 |
+
- Macro avg: {'precision': 0.9449484735388226, 'recall': 0.9548175850599052, 'f1-score': 0.9498022538988083, 'support': 29261.0}
|
| 42 |
+
- Weighted avg: {'precision': 0.9741840360302777, 'recall': 0.9739926865110556, 'f1-score': 0.9740652601681857, 'support': 29261.0}
|
| 43 |
|
| 44 |
## Model description
|
| 45 |
|
|
|
|
| 64 |
- seed: 42
|
| 65 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 66 |
- lr_scheduler_type: linear
|
| 67 |
+
- num_epochs: 5
|
| 68 |
|
| 69 |
### Training results
|
| 70 |
|
| 71 |
| Training Loss | Epoch | Step | Validation Loss | B | I | O | Accuracy | Macro avg | Weighted avg |
|
| 72 |
|:-------------:|:-----:|:----:|:---------------:|:------------------------------------------------------------------------------------------------------------------:|:-------------------------------------------------------------------------------------------------------------------:|:------------------------------------------------------------------------------------------------------------------:|:--------:|:-------------------------------------------------------------------------------------------------------------------:|:-------------------------------------------------------------------------------------------------------------------:|
|
| 73 |
+
| No log | 1.0 | 41 | 0.2075 | {'precision': 0.8258258258258259, 'recall': 0.7281553398058253, 'f1-score': 0.773921200750469, 'support': 1133.0} | {'precision': 0.9305934158104424, 'recall': 0.9712753734201456, 'f1-score': 0.9504992905522983, 'support': 18277.0} | {'precision': 0.9388199433921185, 'recall': 0.8754441173484926, 'f1-score': 0.9060251089982665, 'support': 9851.0} | 0.9296 | {'precision': 0.8984130616761289, 'recall': 0.8582916101914878, 'f1-score': 0.8768152001003445, 'support': 29261.0} | {'precision': 0.9293063047668868, 'recall': 0.9295991251153413, 'f1-score': 0.9286894365406706, 'support': 29261.0} |
|
| 74 |
+
| No log | 2.0 | 82 | 0.1039 | {'precision': 0.7817781043350478, 'recall': 0.93909973521624, 'f1-score': 0.8532477947072975, 'support': 1133.0} | {'precision': 0.9750846901977925, 'recall': 0.9764184494173004, 'f1-score': 0.9757511140271741, 'support': 18277.0} | {'precision': 0.9661387789122734, 'recall': 0.9413257537305857, 'f1-score': 0.9535708776800864, 'support': 9851.0} | 0.9632 | {'precision': 0.9076671911483712, 'recall': 0.9522813127880422, 'f1-score': 0.927523262138186, 'support': 29261.0} | {'precision': 0.9645880382085871, 'recall': 0.9631591538224941, 'f1-score': 0.9635405344487392, 'support': 29261.0} |
|
| 75 |
+
| No log | 3.0 | 123 | 0.0875 | {'precision': 0.8751054852320675, 'recall': 0.9152691968225949, 'f1-score': 0.8947368421052632, 'support': 1133.0} | {'precision': 0.9870288248337029, 'recall': 0.9742299064397877, 'f1-score': 0.9805876036016191, 'support': 18277.0} | {'precision': 0.9561578318055002, 'recall': 0.9741143031164349, 'f1-score': 0.9650525468899281, 'support': 9851.0} | 0.9719 | {'precision': 0.9394307139570902, 'recall': 0.9545378021262724, 'f1-score': 0.9467923308656035, 'support': 29261.0} | {'precision': 0.972302079469926, 'recall': 0.9719080004101022, 'f1-score': 0.9720333929990342, 'support': 29261.0} |
|
| 76 |
+
| No log | 4.0 | 164 | 0.0825 | {'precision': 0.8817021276595745, 'recall': 0.9143865842894969, 'f1-score': 0.8977469670710572, 'support': 1133.0} | {'precision': 0.9845409033393849, 'recall': 0.9791541281391913, 'f1-score': 0.9818401272837, 'support': 18277.0} | {'precision': 0.9638712281764052, 'recall': 0.9695462389605116, 'f1-score': 0.9667004048582996, 'support': 9851.0} | 0.9734 | {'precision': 0.9433714197251216, 'recall': 0.9543623171297333, 'f1-score': 0.9487624997376857, 'support': 29261.0} | {'precision': 0.9736002894548376, 'recall': 0.9734117084173474, 'f1-score': 0.9734870649777793, 'support': 29261.0} |
|
| 77 |
+
| No log | 5.0 | 205 | 0.0856 | {'precision': 0.8861301369863014, 'recall': 0.913503971756399, 'f1-score': 0.8996088657105606, 'support': 1133.0} | {'precision': 0.9856182499448976, 'recall': 0.978661705969251, 'f1-score': 0.9821276595744681, 'support': 18277.0} | {'precision': 0.963097033685269, 'recall': 0.9722870774540656, 'f1-score': 0.9676702364113963, 'support': 9851.0} | 0.9740 | {'precision': 0.9449484735388226, 'recall': 0.9548175850599052, 'f1-score': 0.9498022538988083, 'support': 29261.0} | {'precision': 0.9741840360302777, 'recall': 0.9739926865110556, 'f1-score': 0.9740652601681857, 'support': 29261.0} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 78 |
|
| 79 |
|
| 80 |
### Framework versions
|
| 81 |
|
| 82 |
+
- Transformers 4.45.2
|
| 83 |
+
- Pytorch 2.5.0+cu124
|
| 84 |
+
- Datasets 2.19.1
|
| 85 |
+
- Tokenizers 0.20.1
|
meta_data/README_s42_e5.md
CHANGED
|
@@ -1,9 +1,11 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
| 2 |
base_model: allenai/longformer-base-4096
|
| 3 |
tags:
|
| 4 |
- generated_from_trainer
|
| 5 |
datasets:
|
| 6 |
-
-
|
| 7 |
metrics:
|
| 8 |
- accuracy
|
| 9 |
model-index:
|
|
@@ -13,15 +15,15 @@ model-index:
|
|
| 13 |
name: Token Classification
|
| 14 |
type: token-classification
|
| 15 |
dataset:
|
| 16 |
-
name:
|
| 17 |
-
type:
|
| 18 |
config: spans
|
| 19 |
-
split: train[
|
| 20 |
args: spans
|
| 21 |
metrics:
|
| 22 |
- name: Accuracy
|
| 23 |
type: accuracy
|
| 24 |
-
value: 0.
|
| 25 |
---
|
| 26 |
|
| 27 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
@@ -29,15 +31,15 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 29 |
|
| 30 |
# longformer-spans
|
| 31 |
|
| 32 |
-
This model is a fine-tuned version of [allenai/longformer-base-4096](https://huggingface.co/allenai/longformer-base-4096) on the
|
| 33 |
It achieves the following results on the evaluation set:
|
| 34 |
-
- Loss: 0.
|
| 35 |
-
- B: {'precision': 0.
|
| 36 |
-
- I: {'precision': 0.
|
| 37 |
-
- O: {'precision': 0.
|
| 38 |
-
- Accuracy: 0.
|
| 39 |
-
- Macro avg: {'precision': 0.
|
| 40 |
-
- Weighted avg: {'precision': 0.
|
| 41 |
|
| 42 |
## Model description
|
| 43 |
|
|
@@ -66,18 +68,18 @@ The following hyperparameters were used during training:
|
|
| 66 |
|
| 67 |
### Training results
|
| 68 |
|
| 69 |
-
| Training Loss | Epoch | Step | Validation Loss | B
|
| 70 |
-
|:-------------:|:-----:|:----:|:---------------:|:------------------------------------------------------------------------------------------------------------------
|
| 71 |
-
| No log | 1.0 | 41 | 0.
|
| 72 |
-
| No log | 2.0 | 82 | 0.
|
| 73 |
-
| No log | 3.0 | 123 | 0.
|
| 74 |
-
| No log | 4.0 | 164 | 0.
|
| 75 |
-
| No log | 5.0 | 205 | 0.
|
| 76 |
|
| 77 |
|
| 78 |
### Framework versions
|
| 79 |
|
| 80 |
-
- Transformers 4.
|
| 81 |
-
- Pytorch 2.
|
| 82 |
-
- Datasets 2.
|
| 83 |
-
- Tokenizers 0.
|
|
|
|
| 1 |
---
|
| 2 |
+
library_name: transformers
|
| 3 |
+
license: apache-2.0
|
| 4 |
base_model: allenai/longformer-base-4096
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
| 7 |
datasets:
|
| 8 |
+
- stab-gurevych-essays
|
| 9 |
metrics:
|
| 10 |
- accuracy
|
| 11 |
model-index:
|
|
|
|
| 15 |
name: Token Classification
|
| 16 |
type: token-classification
|
| 17 |
dataset:
|
| 18 |
+
name: stab-gurevych-essays
|
| 19 |
+
type: stab-gurevych-essays
|
| 20 |
config: spans
|
| 21 |
+
split: train[0%:20%]
|
| 22 |
args: spans
|
| 23 |
metrics:
|
| 24 |
- name: Accuracy
|
| 25 |
type: accuracy
|
| 26 |
+
value: 0.9739926865110556
|
| 27 |
---
|
| 28 |
|
| 29 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
|
| 31 |
|
| 32 |
# longformer-spans
|
| 33 |
|
| 34 |
+
This model is a fine-tuned version of [allenai/longformer-base-4096](https://huggingface.co/allenai/longformer-base-4096) on the stab-gurevych-essays dataset.
|
| 35 |
It achieves the following results on the evaluation set:
|
| 36 |
+
- Loss: 0.0856
|
| 37 |
+
- B: {'precision': 0.8861301369863014, 'recall': 0.913503971756399, 'f1-score': 0.8996088657105606, 'support': 1133.0}
|
| 38 |
+
- I: {'precision': 0.9856182499448976, 'recall': 0.978661705969251, 'f1-score': 0.9821276595744681, 'support': 18277.0}
|
| 39 |
+
- O: {'precision': 0.963097033685269, 'recall': 0.9722870774540656, 'f1-score': 0.9676702364113963, 'support': 9851.0}
|
| 40 |
+
- Accuracy: 0.9740
|
| 41 |
+
- Macro avg: {'precision': 0.9449484735388226, 'recall': 0.9548175850599052, 'f1-score': 0.9498022538988083, 'support': 29261.0}
|
| 42 |
+
- Weighted avg: {'precision': 0.9741840360302777, 'recall': 0.9739926865110556, 'f1-score': 0.9740652601681857, 'support': 29261.0}
|
| 43 |
|
| 44 |
## Model description
|
| 45 |
|
|
|
|
| 68 |
|
| 69 |
### Training results
|
| 70 |
|
| 71 |
+
| Training Loss | Epoch | Step | Validation Loss | B | I | O | Accuracy | Macro avg | Weighted avg |
|
| 72 |
+
|:-------------:|:-----:|:----:|:---------------:|:------------------------------------------------------------------------------------------------------------------:|:-------------------------------------------------------------------------------------------------------------------:|:------------------------------------------------------------------------------------------------------------------:|:--------:|:-------------------------------------------------------------------------------------------------------------------:|:-------------------------------------------------------------------------------------------------------------------:|
|
| 73 |
+
| No log | 1.0 | 41 | 0.2075 | {'precision': 0.8258258258258259, 'recall': 0.7281553398058253, 'f1-score': 0.773921200750469, 'support': 1133.0} | {'precision': 0.9305934158104424, 'recall': 0.9712753734201456, 'f1-score': 0.9504992905522983, 'support': 18277.0} | {'precision': 0.9388199433921185, 'recall': 0.8754441173484926, 'f1-score': 0.9060251089982665, 'support': 9851.0} | 0.9296 | {'precision': 0.8984130616761289, 'recall': 0.8582916101914878, 'f1-score': 0.8768152001003445, 'support': 29261.0} | {'precision': 0.9293063047668868, 'recall': 0.9295991251153413, 'f1-score': 0.9286894365406706, 'support': 29261.0} |
|
| 74 |
+
| No log | 2.0 | 82 | 0.1039 | {'precision': 0.7817781043350478, 'recall': 0.93909973521624, 'f1-score': 0.8532477947072975, 'support': 1133.0} | {'precision': 0.9750846901977925, 'recall': 0.9764184494173004, 'f1-score': 0.9757511140271741, 'support': 18277.0} | {'precision': 0.9661387789122734, 'recall': 0.9413257537305857, 'f1-score': 0.9535708776800864, 'support': 9851.0} | 0.9632 | {'precision': 0.9076671911483712, 'recall': 0.9522813127880422, 'f1-score': 0.927523262138186, 'support': 29261.0} | {'precision': 0.9645880382085871, 'recall': 0.9631591538224941, 'f1-score': 0.9635405344487392, 'support': 29261.0} |
|
| 75 |
+
| No log | 3.0 | 123 | 0.0875 | {'precision': 0.8751054852320675, 'recall': 0.9152691968225949, 'f1-score': 0.8947368421052632, 'support': 1133.0} | {'precision': 0.9870288248337029, 'recall': 0.9742299064397877, 'f1-score': 0.9805876036016191, 'support': 18277.0} | {'precision': 0.9561578318055002, 'recall': 0.9741143031164349, 'f1-score': 0.9650525468899281, 'support': 9851.0} | 0.9719 | {'precision': 0.9394307139570902, 'recall': 0.9545378021262724, 'f1-score': 0.9467923308656035, 'support': 29261.0} | {'precision': 0.972302079469926, 'recall': 0.9719080004101022, 'f1-score': 0.9720333929990342, 'support': 29261.0} |
|
| 76 |
+
| No log | 4.0 | 164 | 0.0825 | {'precision': 0.8817021276595745, 'recall': 0.9143865842894969, 'f1-score': 0.8977469670710572, 'support': 1133.0} | {'precision': 0.9845409033393849, 'recall': 0.9791541281391913, 'f1-score': 0.9818401272837, 'support': 18277.0} | {'precision': 0.9638712281764052, 'recall': 0.9695462389605116, 'f1-score': 0.9667004048582996, 'support': 9851.0} | 0.9734 | {'precision': 0.9433714197251216, 'recall': 0.9543623171297333, 'f1-score': 0.9487624997376857, 'support': 29261.0} | {'precision': 0.9736002894548376, 'recall': 0.9734117084173474, 'f1-score': 0.9734870649777793, 'support': 29261.0} |
|
| 77 |
+
| No log | 5.0 | 205 | 0.0856 | {'precision': 0.8861301369863014, 'recall': 0.913503971756399, 'f1-score': 0.8996088657105606, 'support': 1133.0} | {'precision': 0.9856182499448976, 'recall': 0.978661705969251, 'f1-score': 0.9821276595744681, 'support': 18277.0} | {'precision': 0.963097033685269, 'recall': 0.9722870774540656, 'f1-score': 0.9676702364113963, 'support': 9851.0} | 0.9740 | {'precision': 0.9449484735388226, 'recall': 0.9548175850599052, 'f1-score': 0.9498022538988083, 'support': 29261.0} | {'precision': 0.9741840360302777, 'recall': 0.9739926865110556, 'f1-score': 0.9740652601681857, 'support': 29261.0} |
|
| 78 |
|
| 79 |
|
| 80 |
### Framework versions
|
| 81 |
|
| 82 |
+
- Transformers 4.45.2
|
| 83 |
+
- Pytorch 2.5.0+cu124
|
| 84 |
+
- Datasets 2.19.1
|
| 85 |
+
- Tokenizers 0.20.1
|
meta_data/meta_s42_e5_cvi0.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"B": {"precision": 0.
|
|
|
|
| 1 |
+
{"B": {"precision": 0.8861301369863014, "recall": 0.913503971756399, "f1-score": 0.8996088657105606, "support": 1133.0}, "I": {"precision": 0.9856182499448976, "recall": 0.978661705969251, "f1-score": 0.9821276595744681, "support": 18277.0}, "O": {"precision": 0.963097033685269, "recall": 0.9722870774540656, "f1-score": 0.9676702364113963, "support": 9851.0}, "accuracy": 0.9739926865110556, "macro avg": {"precision": 0.9449484735388226, "recall": 0.9548175850599052, "f1-score": 0.9498022538988083, "support": 29261.0}, "weighted avg": {"precision": 0.9741840360302777, "recall": 0.9739926865110556, "f1-score": 0.9740652601681857, "support": 29261.0}}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 592318676
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41b8d40517218dac82dd91264578c07f8eca90ed92129f696683f1b8e080f4a2
|
| 3 |
size 592318676
|