Training complete
Browse files- README.md +53 -0
- test_metrics.json +4 -0
- train_losses.csv +112 -0
README.md
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
base_model: FacebookAI/roberta-base
|
| 4 |
+
tags:
|
| 5 |
+
- generated_from_trainer
|
| 6 |
+
model-index:
|
| 7 |
+
- name: roberta_base_QA_SQUAD
|
| 8 |
+
results: []
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 12 |
+
should probably proofread and complete it, then remove this comment. -->
|
| 13 |
+
|
| 14 |
+
# roberta_base_QA_SQUAD
|
| 15 |
+
|
| 16 |
+
This model is a fine-tuned version of [FacebookAI/roberta-base](https://huggingface.co/FacebookAI/roberta-base) on an unknown dataset.
|
| 17 |
+
|
| 18 |
+
## Model description
|
| 19 |
+
|
| 20 |
+
More information needed
|
| 21 |
+
|
| 22 |
+
## Intended uses & limitations
|
| 23 |
+
|
| 24 |
+
More information needed
|
| 25 |
+
|
| 26 |
+
## Training and evaluation data
|
| 27 |
+
|
| 28 |
+
More information needed
|
| 29 |
+
|
| 30 |
+
## Training procedure
|
| 31 |
+
|
| 32 |
+
### Training hyperparameters
|
| 33 |
+
|
| 34 |
+
The following hyperparameters were used during training:
|
| 35 |
+
- learning_rate: 2e-05
|
| 36 |
+
- train_batch_size: 8
|
| 37 |
+
- eval_batch_size: 8
|
| 38 |
+
- seed: 42
|
| 39 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 40 |
+
- lr_scheduler_type: linear
|
| 41 |
+
- num_epochs: 5
|
| 42 |
+
- mixed_precision_training: Native AMP
|
| 43 |
+
|
| 44 |
+
### Training results
|
| 45 |
+
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
### Framework versions
|
| 49 |
+
|
| 50 |
+
- Transformers 4.40.0
|
| 51 |
+
- Pytorch 2.2.1
|
| 52 |
+
- Datasets 2.18.0
|
| 53 |
+
- Tokenizers 0.19.1
|
test_metrics.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"exact_match": 84.65468306527909,
|
| 3 |
+
"f1": 91.55804062843333
|
| 4 |
+
}
|
train_losses.csv
ADDED
|
@@ -0,0 +1,112 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
loss,epoch
|
| 2 |
+
2.073,0.04516303856923494
|
| 3 |
+
1.3268,0.09032607713846988
|
| 4 |
+
1.2112,0.1354891157077048
|
| 5 |
+
1.1183,0.18065215427693976
|
| 6 |
+
1.0995,0.22581519284617468
|
| 7 |
+
1.0734,0.2709782314154096
|
| 8 |
+
1.0258,0.3161412699846446
|
| 9 |
+
1.0602,0.3613043085538795
|
| 10 |
+
1.0067,0.40646734712311444
|
| 11 |
+
0.9985,0.45163038569234937
|
| 12 |
+
0.9358,0.4967934242615843
|
| 13 |
+
1.0052,0.5419564628308192
|
| 14 |
+
0.9595,0.5871195014000542
|
| 15 |
+
0.9575,0.6322825399692892
|
| 16 |
+
0.9399,0.677445578538524
|
| 17 |
+
0.9114,0.722608617107759
|
| 18 |
+
0.9735,0.7677716556769939
|
| 19 |
+
0.9248,0.8129346942462289
|
| 20 |
+
0.9132,0.8580977328154639
|
| 21 |
+
0.9043,0.9032607713846987
|
| 22 |
+
0.9073,0.9484238099539337
|
| 23 |
+
0.908,0.9935868485231686
|
| 24 |
+
0.7482,1.0387498870924037
|
| 25 |
+
0.7074,1.0839129256616384
|
| 26 |
+
0.702,1.1290759642308734
|
| 27 |
+
0.7315,1.1742390028001084
|
| 28 |
+
0.6674,1.2194020413693434
|
| 29 |
+
0.759,1.2645650799385781
|
| 30 |
+
0.7187,1.3097281185078131
|
| 31 |
+
0.6939,1.354891157077048
|
| 32 |
+
0.7177,1.400054195646283
|
| 33 |
+
0.7284,1.445217234215518
|
| 34 |
+
0.6842,1.490380272784753
|
| 35 |
+
0.7188,1.535543311353988
|
| 36 |
+
0.7052,1.5807063499232228
|
| 37 |
+
0.7088,1.6258693884924578
|
| 38 |
+
0.7039,1.6710324270616927
|
| 39 |
+
0.7191,1.7161954656309275
|
| 40 |
+
0.7319,1.7613585042001625
|
| 41 |
+
0.6991,1.8065215427693975
|
| 42 |
+
0.7152,1.8516845813386325
|
| 43 |
+
0.7149,1.8968476199078674
|
| 44 |
+
0.7272,1.9420106584771024
|
| 45 |
+
0.6842,1.9871736970463374
|
| 46 |
+
0.5655,2.0323367356155724
|
| 47 |
+
0.5356,2.0774997741848074
|
| 48 |
+
0.5181,2.122662812754042
|
| 49 |
+
0.5664,2.167825851323277
|
| 50 |
+
0.5077,2.212988889892512
|
| 51 |
+
0.5628,2.258151928461747
|
| 52 |
+
0.5437,2.303314967030982
|
| 53 |
+
0.5575,2.348478005600217
|
| 54 |
+
0.5216,2.393641044169452
|
| 55 |
+
0.522,2.4388040827386868
|
| 56 |
+
0.5374,2.4839671213079217
|
| 57 |
+
0.538,2.5291301598771563
|
| 58 |
+
0.5293,2.5742931984463917
|
| 59 |
+
0.549,2.6194562370156262
|
| 60 |
+
0.5579,2.6646192755848612
|
| 61 |
+
0.5228,2.709782314154096
|
| 62 |
+
0.5352,2.754945352723331
|
| 63 |
+
0.5427,2.800108391292566
|
| 64 |
+
0.5062,2.845271429861801
|
| 65 |
+
0.5242,2.890434468431036
|
| 66 |
+
0.5492,2.935597507000271
|
| 67 |
+
0.5455,2.980760545569506
|
| 68 |
+
0.4557,3.025923584138741
|
| 69 |
+
0.4097,3.0710866227079756
|
| 70 |
+
0.4264,3.1162496612772106
|
| 71 |
+
0.4034,3.1614126998464456
|
| 72 |
+
0.4014,3.2065757384156806
|
| 73 |
+
0.4157,3.2517387769849155
|
| 74 |
+
0.417,3.2969018155541505
|
| 75 |
+
0.4055,3.3420648541233855
|
| 76 |
+
0.4273,3.3872278926926205
|
| 77 |
+
0.4075,3.4323909312618555
|
| 78 |
+
0.3917,3.4775539698310904
|
| 79 |
+
0.4171,3.522717008400325
|
| 80 |
+
0.4211,3.56788004696956
|
| 81 |
+
0.4009,3.613043085538795
|
| 82 |
+
0.3833,3.65820612410803
|
| 83 |
+
0.3732,3.703369162677265
|
| 84 |
+
0.4157,3.7485322012465
|
| 85 |
+
0.3983,3.793695239815735
|
| 86 |
+
0.3876,3.83885827838497
|
| 87 |
+
0.3889,3.884021316954205
|
| 88 |
+
0.4027,3.9291843555234394
|
| 89 |
+
0.3962,3.974347394092675
|
| 90 |
+
0.3449,4.019510432661909
|
| 91 |
+
0.3182,4.064673471231145
|
| 92 |
+
0.302,4.109836509800379
|
| 93 |
+
0.3066,4.154999548369615
|
| 94 |
+
0.3204,4.200162586938849
|
| 95 |
+
0.3105,4.245325625508084
|
| 96 |
+
0.3105,4.290488664077319
|
| 97 |
+
0.3184,4.335651702646554
|
| 98 |
+
0.3306,4.380814741215789
|
| 99 |
+
0.3136,4.425977779785024
|
| 100 |
+
0.3074,4.471140818354259
|
| 101 |
+
0.3458,4.516303856923494
|
| 102 |
+
0.302,4.561466895492729
|
| 103 |
+
0.3051,4.606629934061964
|
| 104 |
+
0.3089,4.651792972631199
|
| 105 |
+
0.2946,4.696956011200434
|
| 106 |
+
0.3046,4.742119049769668
|
| 107 |
+
0.3206,4.787282088338904
|
| 108 |
+
0.2986,4.832445126908138
|
| 109 |
+
0.2949,4.8776081654773735
|
| 110 |
+
0.2964,4.922771204046608
|
| 111 |
+
0.2939,4.9679342426158435
|
| 112 |
+
0.6032335646784992,5.0
|