igerasimov/nasa-impact-bert-e-base-mlm-finetuned
Browse files- README.md +173 -0
- config.json +78 -0
- model.safetensors +3 -0
- training_args.bin +3 -0
README.md
ADDED
|
@@ -0,0 +1,173 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
base_model: nasa-impact/nasa-smd-ibm-st
|
| 4 |
+
tags:
|
| 5 |
+
- generated_from_trainer
|
| 6 |
+
metrics:
|
| 7 |
+
- accuracy
|
| 8 |
+
model-index:
|
| 9 |
+
- name: checkpoints
|
| 10 |
+
results: []
|
| 11 |
+
---
|
| 12 |
+
|
| 13 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 14 |
+
should probably proofread and complete it, then remove this comment. -->
|
| 15 |
+
|
| 16 |
+
# checkpoints
|
| 17 |
+
|
| 18 |
+
This model is a fine-tuned version of [nasa-impact/nasa-smd-ibm-st](https://huggingface.co/nasa-impact/nasa-smd-ibm-st) on the None dataset.
|
| 19 |
+
It achieves the following results on the evaluation set:
|
| 20 |
+
- Accuracy: 0.7087
|
| 21 |
+
- Loss: 1.5329
|
| 22 |
+
- Overall/weighted Precision: 0.7311
|
| 23 |
+
- Overall/weighted Recall: 0.7087
|
| 24 |
+
- Overall/weighted F1: 0.7105
|
| 25 |
+
- Class Agriculture/precision: 1.0
|
| 26 |
+
- Class Agriculture/recall: 1.0
|
| 27 |
+
- Class Agriculture/f1: 1.0
|
| 28 |
+
- Class Air quality/precision: 0.7586
|
| 29 |
+
- Class Air quality/recall: 0.8462
|
| 30 |
+
- Class Air quality/f1: 0.8
|
| 31 |
+
- Class Cryospheric climate indicators/precision: 1.0
|
| 32 |
+
- Class Cryospheric climate indicators/recall: 0.5
|
| 33 |
+
- Class Cryospheric climate indicators/f1: 0.6667
|
| 34 |
+
- Class Droughts/precision: 1.0
|
| 35 |
+
- Class Droughts/recall: 1.0
|
| 36 |
+
- Class Droughts/f1: 1.0
|
| 37 |
+
- Class Earthquakes/precision: 1.0
|
| 38 |
+
- Class Earthquakes/recall: 1.0
|
| 39 |
+
- Class Earthquakes/f1: 1.0
|
| 40 |
+
- Class Ecosystem species/precision: 0.8571
|
| 41 |
+
- Class Ecosystem species/recall: 0.75
|
| 42 |
+
- Class Ecosystem species/f1: 0.8
|
| 43 |
+
- Class Ecosystems/precision: 0.0
|
| 44 |
+
- Class Ecosystems/recall: 1.0
|
| 45 |
+
- Class Ecosystems/f1: 0.0
|
| 46 |
+
- Class Energy production/use/precision: 0.0
|
| 47 |
+
- Class Energy production/use/recall: 0.0
|
| 48 |
+
- Class Energy production/use/f1: 0.0
|
| 49 |
+
- Class Extreme weather/precision: 0.3846
|
| 50 |
+
- Class Extreme weather/recall: 0.8333
|
| 51 |
+
- Class Extreme weather/f1: 0.5263
|
| 52 |
+
- Class Floods/precision: 0.0
|
| 53 |
+
- Class Floods/recall: 0.0
|
| 54 |
+
- Class Floods/f1: 0.0
|
| 55 |
+
- Class Greenhouse gases/precision: 0.6667
|
| 56 |
+
- Class Greenhouse gases/recall: 0.5
|
| 57 |
+
- Class Greenhouse gases/f1: 0.5714
|
| 58 |
+
- Class Heat/precision: 0.75
|
| 59 |
+
- Class Heat/recall: 0.75
|
| 60 |
+
- Class Heat/f1: 0.75
|
| 61 |
+
- Class Land use and cover change/precision: 1.0
|
| 62 |
+
- Class Land use and cover change/recall: 1.0
|
| 63 |
+
- Class Land use and cover change/f1: 1.0
|
| 64 |
+
- Class Landslides/precision: 0.8333
|
| 65 |
+
- Class Landslides/recall: 0.8333
|
| 66 |
+
- Class Landslides/f1: 0.8333
|
| 67 |
+
- Class Public health/precision: 1.0
|
| 68 |
+
- Class Public health/recall: 0.5
|
| 69 |
+
- Class Public health/f1: 0.6667
|
| 70 |
+
- Class Severe storms/precision: 0.5
|
| 71 |
+
- Class Severe storms/recall: 0.4
|
| 72 |
+
- Class Severe storms/f1: 0.4444
|
| 73 |
+
- Class Sun-earth interactions/precision: 0.7667
|
| 74 |
+
- Class Sun-earth interactions/recall: 0.6970
|
| 75 |
+
- Class Sun-earth interactions/f1: 0.7302
|
| 76 |
+
- Class Teleconnections/precision: 0.6667
|
| 77 |
+
- Class Teleconnections/recall: 0.5455
|
| 78 |
+
- Class Teleconnections/f1: 0.6
|
| 79 |
+
- Class Temperature indicators/precision: 0.5
|
| 80 |
+
- Class Temperature indicators/recall: 0.6667
|
| 81 |
+
- Class Temperature indicators/f1: 0.5714
|
| 82 |
+
|
| 83 |
+
## Model description
|
| 84 |
+
|
| 85 |
+
More information needed
|
| 86 |
+
|
| 87 |
+
## Intended uses & limitations
|
| 88 |
+
|
| 89 |
+
More information needed
|
| 90 |
+
|
| 91 |
+
## Training and evaluation data
|
| 92 |
+
|
| 93 |
+
More information needed
|
| 94 |
+
|
| 95 |
+
## Training procedure
|
| 96 |
+
|
| 97 |
+
### Training hyperparameters
|
| 98 |
+
|
| 99 |
+
The following hyperparameters were used during training:
|
| 100 |
+
- learning_rate: 5e-05
|
| 101 |
+
- train_batch_size: 16
|
| 102 |
+
- eval_batch_size: 16
|
| 103 |
+
- seed: 42
|
| 104 |
+
- gradient_accumulation_steps: 4
|
| 105 |
+
- total_train_batch_size: 64
|
| 106 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 107 |
+
- lr_scheduler_type: linear
|
| 108 |
+
- lr_scheduler_warmup_steps: 500
|
| 109 |
+
- num_epochs: 50
|
| 110 |
+
- mixed_precision_training: Native AMP
|
| 111 |
+
|
| 112 |
+
### Training results
|
| 113 |
+
|
| 114 |
+
| Training Loss | Epoch | Step | Accuracy | Validation Loss | Overall/weighted Precision | Overall/weighted Recall | Overall/weighted F1 | Class Agriculture/precision | Class Agriculture/recall | Class Agriculture/f1 | Class Air quality/precision | Class Air quality/recall | Class Air quality/f1 | Class Cryospheric climate indicators/precision | Class Cryospheric climate indicators/recall | Class Cryospheric climate indicators/f1 | Class Droughts/precision | Class Droughts/recall | Class Droughts/f1 | Class Earthquakes/precision | Class Earthquakes/recall | Class Earthquakes/f1 | Class Ecosystem species/precision | Class Ecosystem species/recall | Class Ecosystem species/f1 | Class Ecosystems/precision | Class Ecosystems/recall | Class Ecosystems/f1 | Class Energy production/use/precision | Class Energy production/use/recall | Class Energy production/use/f1 | Class Extreme weather/precision | Class Extreme weather/recall | Class Extreme weather/f1 | Class Floods/precision | Class Floods/recall | Class Floods/f1 | Class Greenhouse gases/precision | Class Greenhouse gases/recall | Class Greenhouse gases/f1 | Class Heat/precision | Class Heat/recall | Class Heat/f1 | Class Land use and cover change/precision | Class Land use and cover change/recall | Class Land use and cover change/f1 | Class Landslides/precision | Class Landslides/recall | Class Landslides/f1 | Class Public health/precision | Class Public health/recall | Class Public health/f1 | Class Severe storms/precision | Class Severe storms/recall | Class Severe storms/f1 | Class Sun-earth interactions/precision | Class Sun-earth interactions/recall | Class Sun-earth interactions/f1 | Class Teleconnections/precision | Class Teleconnections/recall | Class Teleconnections/f1 | Class Temperature indicators/precision | Class Temperature indicators/recall | Class Temperature indicators/f1 | Class Validation/precision | Class Validation/recall | Class Validation/f1 |
|
| 115 |
+
|:-------------:|:-----:|:----:|:--------:|:---------------:|:--------------------------:|:-----------------------:|:-------------------:|:---------------------------:|:------------------------:|:--------------------:|:---------------------------:|:------------------------:|:--------------------:|:----------------------------------------------:|:-------------------------------------------:|:---------------------------------------:|:------------------------:|:---------------------:|:-----------------:|:---------------------------:|:------------------------:|:--------------------:|:---------------------------------:|:------------------------------:|:--------------------------:|:--------------------------:|:-----------------------:|:-------------------:|:-------------------------------------:|:----------------------------------:|:------------------------------:|:-------------------------------:|:----------------------------:|:------------------------:|:----------------------:|:-------------------:|:---------------:|:--------------------------------:|:-----------------------------:|:-------------------------:|:--------------------:|:-----------------:|:-------------:|:-----------------------------------------:|:--------------------------------------:|:----------------------------------:|:--------------------------:|:-----------------------:|:-------------------:|:-----------------------------:|:--------------------------:|:----------------------:|:-----------------------------:|:--------------------------:|:----------------------:|:--------------------------------------:|:-----------------------------------:|:-------------------------------:|:-------------------------------:|:----------------------------:|:------------------------:|:--------------------------------------:|:-----------------------------------:|:-------------------------------:|:--------------------------:|:-----------------------:|:-------------------:|
|
| 116 |
+
| No log | 1.0 | 8 | 0.0 | 3.1300 | 0.4173 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 117 |
+
| 3.1343 | 2.0 | 16 | 0.0945 | 3.0987 | 0.4665 | 0.0945 | 0.0646 | 1.0 | 0.0 | 0.0 | 0.24 | 0.4615 | 0.3158 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 118 |
+
| 3.1018 | 3.0 | 24 | 0.2126 | 3.0484 | 0.5508 | 0.2126 | 0.0978 | 1.0 | 0.0 | 0.0 | 0.2336 | 0.9615 | 0.3759 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.1176 | 0.0606 | 0.08 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 119 |
+
| 3.0499 | 4.0 | 32 | 0.2283 | 2.9848 | 0.6297 | 0.2283 | 0.1095 | 1.0 | 0.0 | 0.0 | 0.2364 | 1.0 | 0.3824 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.1765 | 0.0909 | 0.12 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 120 |
+
| 2.9715 | 5.0 | 40 | 0.2283 | 2.9113 | 0.6321 | 0.2283 | 0.1095 | 1.0 | 0.0 | 0.0 | 0.2342 | 1.0 | 0.3796 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.1875 | 0.0909 | 0.1224 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 121 |
+
| 2.9715 | 6.0 | 48 | 0.2520 | 2.8317 | 0.6668 | 0.2520 | 0.1394 | 1.0 | 0.0 | 0.0 | 0.2407 | 1.0 | 0.3881 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3158 | 0.1818 | 0.2308 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 122 |
+
| 2.8735 | 7.0 | 56 | 0.4016 | 2.7478 | 0.7263 | 0.4016 | 0.2582 | 1.0 | 0.0 | 0.0 | 0.52 | 1.0 | 0.6842 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3247 | 0.7576 | 0.4545 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 123 |
+
| 2.7766 | 8.0 | 64 | 0.3937 | 2.6696 | 0.7255 | 0.3937 | 0.2552 | 1.0 | 0.0 | 0.0 | 0.5319 | 0.9615 | 0.6849 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3125 | 0.7576 | 0.4425 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 124 |
+
| 2.6778 | 9.0 | 72 | 0.3307 | 2.6007 | 0.7783 | 0.3307 | 0.2201 | 1.0 | 0.0 | 0.0 | 0.8333 | 0.3846 | 0.5263 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.2783 | 0.9697 | 0.4324 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 125 |
+
| 2.6132 | 10.0 | 80 | 0.3701 | 2.5347 | 0.7679 | 0.3701 | 0.2580 | 1.0 | 0.0 | 0.0 | 0.7727 | 0.6538 | 0.7083 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.2857 | 0.9091 | 0.4348 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 126 |
+
| 2.6132 | 11.0 | 88 | 0.3937 | 2.4538 | 0.7650 | 0.3937 | 0.2717 | 1.0 | 0.0 | 0.0 | 0.7407 | 0.7692 | 0.7547 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3 | 0.9091 | 0.4511 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 127 |
+
| 2.502 | 12.0 | 96 | 0.4252 | 2.3546 | 0.7537 | 0.4252 | 0.3030 | 1.0 | 0.0 | 0.0 | 0.6667 | 0.9231 | 0.7742 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.25 | 0.4 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3146 | 0.8485 | 0.4590 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 128 |
+
| 2.4298 | 13.0 | 104 | 0.4646 | 2.2607 | 0.7308 | 0.4646 | 0.3283 | 1.0 | 0.0 | 0.0 | 0.6098 | 0.9615 | 0.7463 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6 | 0.75 | 0.6667 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3684 | 0.8485 | 0.5138 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 |
|
| 129 |
+
| 2.263 | 14.0 | 112 | 0.4803 | 2.1656 | 0.6757 | 0.4803 | 0.3474 | 1.0 | 0.0 | 0.0 | 0.6098 | 0.9615 | 0.7463 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.4375 | 0.875 | 0.5833 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.4179 | 0.8485 | 0.56 | 0.3333 | 0.0909 | 0.1429 | 1.0 | 0.0 | 0.0 |
|
| 130 |
+
| 2.1649 | 15.0 | 120 | 0.4882 | 2.0712 | 0.6239 | 0.4882 | 0.3616 | 1.0 | 0.0 | 0.0 | 0.5814 | 0.9615 | 0.7246 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.5 | 0.75 | 0.6 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.4462 | 0.8788 | 0.5918 | 0.4 | 0.1818 | 0.25 | 1.0 | 0.0 | 0.0 |
|
| 131 |
+
| 2.1649 | 16.0 | 128 | 0.5118 | 1.9789 | 0.6575 | 0.5118 | 0.3960 | 1.0 | 0.0 | 0.0 | 0.6944 | 0.9615 | 0.8065 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.4615 | 0.75 | 0.5714 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.1538 | 0.25 | 0.1905 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.5082 | 0.9394 | 0.6596 | 0.25 | 0.0909 | 0.1333 | 1.0 | 0.0 | 0.0 |
|
| 132 |
+
| 2.0169 | 17.0 | 136 | 0.5433 | 1.9319 | 0.7150 | 0.5433 | 0.4470 | 1.0 | 0.0 | 0.0 | 0.6757 | 0.9615 | 0.7937 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.4286 | 0.75 | 0.5455 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.1818 | 0.5 | 0.2667 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6383 | 0.9091 | 0.75 | 0.5714 | 0.3636 | 0.4444 | 1.0 | 0.0 | 0.0 |
|
| 133 |
+
| 1.8937 | 18.0 | 144 | 0.5276 | 1.8886 | 0.7051 | 0.5276 | 0.4316 | 1.0 | 0.0 | 0.0 | 0.5814 | 0.9615 | 0.7246 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.4615 | 0.75 | 0.5714 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.25 | 0.75 | 0.375 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.65 | 0.7879 | 0.7123 | 0.5714 | 0.3636 | 0.4444 | 1.0 | 0.0 | 0.0 |
|
| 134 |
+
| 1.8209 | 19.0 | 152 | 0.5512 | 1.7950 | 0.7211 | 0.5512 | 0.4524 | 1.0 | 0.0 | 0.0 | 0.6579 | 0.9615 | 0.7812 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.5 | 0.75 | 0.6 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.24 | 0.75 | 0.3636 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6444 | 0.8788 | 0.7436 | 0.5714 | 0.3636 | 0.4444 | 1.0 | 0.0 | 0.0 |
|
| 135 |
+
| 1.6452 | 20.0 | 160 | 0.5827 | 1.7998 | 0.7483 | 0.5827 | 0.4947 | 1.0 | 0.0 | 0.0 | 0.625 | 0.9615 | 0.7576 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6667 | 0.75 | 0.7059 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.24 | 0.75 | 0.3636 | 1.0 | 0.25 | 0.4 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6829 | 0.8485 | 0.7568 | 0.7273 | 0.7273 | 0.7273 | 1.0 | 0.0 | 0.0 |
|
| 136 |
+
| 1.6452 | 21.0 | 168 | 0.5827 | 1.7016 | 0.7211 | 0.5827 | 0.5068 | 1.0 | 0.0 | 0.0 | 0.6757 | 0.9615 | 0.7937 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.5833 | 0.875 | 0.7 | 1.0 | 0.0 | 0.0 | 0.5 | 0.1667 | 0.25 | 1.0 | 0.0 | 0.0 | 0.25 | 0.75 | 0.375 | 1.0 | 0.25 | 0.4 | 1.0 | 0.25 | 0.4 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6512 | 0.8485 | 0.7368 | 0.7143 | 0.4545 | 0.5556 | 1.0 | 0.0 | 0.0 |
|
| 137 |
+
| 1.5597 | 22.0 | 176 | 0.6378 | 1.6544 | 0.7338 | 0.6378 | 0.5642 | 1.0 | 0.0 | 0.0 | 0.6579 | 0.9615 | 0.7812 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.5 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.6 | 0.5 | 0.5455 | 1.0 | 0.0 | 0.0 | 0.4118 | 0.875 | 0.56 | 0.75 | 0.75 | 0.75 | 1.0 | 0.25 | 0.4 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6512 | 0.8485 | 0.7368 | 0.7 | 0.6364 | 0.6667 | 1.0 | 0.0 | 0.0 |
|
| 138 |
+
| 1.4316 | 23.0 | 184 | 0.6850 | 1.5807 | 0.7513 | 0.6850 | 0.6319 | 1.0 | 0.0 | 0.0 | 0.7742 | 0.9231 | 0.8421 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.5 | 0.6667 | 0.6667 | 0.75 | 0.7059 | 1.0 | 0.0 | 0.0 | 0.5 | 0.5 | 0.5 | 1.0 | 0.0 | 0.0 | 0.4 | 0.75 | 0.5217 | 0.6 | 0.75 | 0.6667 | 1.0 | 0.75 | 0.8571 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6977 | 0.9091 | 0.7895 | 0.6667 | 0.5455 | 0.6 | 1.0 | 0.0 | 0.0 |
|
| 139 |
+
| 1.2604 | 24.0 | 192 | 0.6850 | 1.5698 | 0.7486 | 0.6850 | 0.6436 | 1.0 | 0.0 | 0.0 | 0.7353 | 0.9615 | 0.8333 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 1.0 | 0.5 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.4286 | 0.5 | 0.4615 | 1.0 | 0.0 | 0.0 | 0.4 | 0.75 | 0.5217 | 0.75 | 0.75 | 0.75 | 1.0 | 0.75 | 0.8571 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.7027 | 0.7879 | 0.7429 | 0.6364 | 0.6364 | 0.6364 | 1.0 | 0.6667 | 0.8 |
|
| 140 |
+
| 1.17 | 25.0 | 200 | 0.6772 | 1.5075 | 0.7379 | 0.6772 | 0.6377 | 1.0 | 0.0 | 0.0 | 0.7353 | 0.9615 | 0.8333 | 1.0 | 0.0 | 0.0 | 1.0 | 0.5 | 0.6667 | 1.0 | 0.5 | 0.6667 | 0.6667 | 0.75 | 0.7059 | 1.0 | 0.0 | 0.0 | 0.4444 | 0.6667 | 0.5333 | 1.0 | 0.0 | 0.0 | 0.4615 | 0.75 | 0.5714 | 0.75 | 0.75 | 0.75 | 1.0 | 0.75 | 0.8571 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6757 | 0.7576 | 0.7143 | 0.6 | 0.5455 | 0.5714 | 1.0 | 0.3333 | 0.5 |
|
| 141 |
+
| 1.17 | 26.0 | 208 | 0.7165 | 1.4437 | 0.7679 | 0.7165 | 0.6756 | 1.0 | 0.0 | 0.0 | 0.8065 | 0.9615 | 0.8772 | 1.0 | 0.0 | 0.0 | 1.0 | 0.5 | 0.6667 | 1.0 | 0.5 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.4444 | 0.6667 | 0.5333 | 1.0 | 0.0 | 0.0 | 0.5455 | 0.75 | 0.6316 | 0.6 | 0.75 | 0.6667 | 1.0 | 0.75 | 0.8571 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6905 | 0.8788 | 0.7733 | 0.6667 | 0.5455 | 0.6 | 1.0 | 0.6667 | 0.8 |
|
| 142 |
+
| 1.0515 | 27.0 | 216 | 0.7402 | 1.3992 | 0.7587 | 0.7402 | 0.7103 | 1.0 | 0.0 | 0.0 | 0.8214 | 0.8846 | 0.8519 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.5 | 0.8333 | 0.625 | 1.0 | 0.0 | 0.0 | 0.8333 | 0.625 | 0.7143 | 1.0 | 0.75 | 0.8571 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.6667 | 0.9091 | 0.7692 | 0.6667 | 0.5455 | 0.6 | 1.0 | 0.6667 | 0.8 |
|
| 143 |
+
| 0.9513 | 28.0 | 224 | 0.7165 | 1.3520 | 0.7659 | 0.7165 | 0.6754 | 1.0 | 0.0 | 0.0 | 0.7667 | 0.8846 | 0.8214 | 1.0 | 0.0 | 0.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.5 | 0.8333 | 0.625 | 1.0 | 0.0 | 0.0 | 0.6667 | 0.75 | 0.7059 | 1.0 | 0.75 | 0.8571 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6667 | 0.8485 | 0.7467 | 0.5455 | 0.5455 | 0.5455 | 1.0 | 0.3333 | 0.5 |
|
| 144 |
+
| 0.7942 | 29.0 | 232 | 0.7244 | 1.3103 | 0.7409 | 0.7244 | 0.7038 | 1.0 | 0.5 | 0.6667 | 0.8065 | 0.9615 | 0.8772 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.6667 | 0.75 | 0.7059 | 1.0 | 0.0 | 0.0 | 0.4167 | 0.8333 | 0.5556 | 1.0 | 0.0 | 0.0 | 0.75 | 0.75 | 0.75 | 1.0 | 0.75 | 0.8571 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.7059 | 0.7273 | 0.7164 | 0.5455 | 0.5455 | 0.5455 | 1.0 | 0.6667 | 0.8 |
|
| 145 |
+
| 0.7386 | 30.0 | 240 | 0.7008 | 1.3628 | 0.7289 | 0.7008 | 0.6813 | 1.0 | 0.0 | 0.0 | 0.8065 | 0.9615 | 0.8772 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.3333 | 0.6667 | 0.4444 | 1.0 | 0.0 | 0.0 | 0.7143 | 0.625 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.7059 | 0.7273 | 0.7164 | 0.6 | 0.5455 | 0.5714 | 0.6667 | 0.6667 | 0.6667 |
|
| 146 |
+
| 0.7386 | 31.0 | 248 | 0.7165 | 1.3002 | 0.7389 | 0.7165 | 0.6980 | 1.0 | 0.5 | 0.6667 | 0.7353 | 0.9615 | 0.8333 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.4 | 0.6667 | 0.5 | 1.0 | 0.0 | 0.0 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.7586 | 0.6667 | 0.7097 | 0.5714 | 0.7273 | 0.64 | 1.0 | 0.6667 | 0.8 |
|
| 147 |
+
| 0.64 | 32.0 | 256 | 0.7244 | 1.3095 | 0.7345 | 0.7244 | 0.7027 | 1.0 | 0.5 | 0.6667 | 0.8065 | 0.9615 | 0.8772 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.4 | 0.6667 | 0.5 | 1.0 | 0.0 | 0.0 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.6944 | 0.7576 | 0.7246 | 0.5455 | 0.5455 | 0.5455 | 1.0 | 0.6667 | 0.8 |
|
| 148 |
+
| 0.5528 | 33.0 | 264 | 0.7087 | 1.2698 | 0.7282 | 0.7087 | 0.6903 | 1.0 | 0.5 | 0.6667 | 0.7576 | 0.9615 | 0.8475 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.3636 | 0.6667 | 0.4706 | 1.0 | 0.0 | 0.0 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.6970 | 0.6970 | 0.6970 | 0.6 | 0.5455 | 0.5714 | 1.0 | 0.6667 | 0.8 |
|
| 149 |
+
| 0.4612 | 34.0 | 272 | 0.7402 | 1.2532 | 0.7488 | 0.7402 | 0.7174 | 1.0 | 1.0 | 1.0 | 0.7812 | 0.9615 | 0.8621 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.4 | 0.6667 | 0.5 | 1.0 | 0.0 | 0.0 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.7027 | 0.7879 | 0.7429 | 0.6667 | 0.5455 | 0.6 | 1.0 | 0.6667 | 0.8 |
|
| 150 |
+
| 0.4051 | 35.0 | 280 | 0.7087 | 1.2745 | 0.7369 | 0.7087 | 0.6930 | 1.0 | 1.0 | 1.0 | 0.7812 | 0.9615 | 0.8621 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3846 | 0.8333 | 0.5263 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.7143 | 0.625 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.7241 | 0.6364 | 0.6774 | 0.5 | 0.6364 | 0.56 | 1.0 | 0.6667 | 0.8 |
|
| 151 |
+
| 0.4051 | 36.0 | 288 | 0.7087 | 1.3092 | 0.7413 | 0.7087 | 0.6988 | 1.0 | 1.0 | 1.0 | 0.8276 | 0.9231 | 0.8727 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3077 | 0.6667 | 0.4211 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.6765 | 0.6970 | 0.6866 | 0.6 | 0.5455 | 0.5714 | 1.0 | 0.6667 | 0.8 |
|
| 152 |
+
| 0.3354 | 37.0 | 296 | 0.6929 | 1.2752 | 0.7348 | 0.6929 | 0.6799 | 1.0 | 0.5 | 0.6667 | 0.7143 | 0.9615 | 0.8197 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.75 | 0.75 | 0.75 | 1.0 | 0.0 | 0.0 | 0.3333 | 0.6667 | 0.4444 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.7143 | 0.625 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.3333 | 0.2 | 0.25 | 0.76 | 0.5758 | 0.6552 | 0.5714 | 0.7273 | 0.64 | 0.6667 | 0.6667 | 0.6667 |
|
| 153 |
+
| 0.2905 | 38.0 | 304 | 0.7244 | 1.2621 | 0.7442 | 0.7244 | 0.7060 | 1.0 | 1.0 | 1.0 | 0.7812 | 0.9615 | 0.8621 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3846 | 0.8333 | 0.5263 | 1.0 | 0.0 | 0.0 | 0.6667 | 0.5 | 0.5714 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.7273 | 0.7273 | 0.7273 | 0.7 | 0.6364 | 0.6667 | 0.6667 | 0.6667 | 0.6667 |
|
| 154 |
+
| 0.249 | 39.0 | 312 | 0.7165 | 1.3443 | 0.7636 | 0.7165 | 0.7027 | 1.0 | 1.0 | 1.0 | 0.7143 | 0.9615 | 0.8197 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3846 | 0.8333 | 0.5263 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.7143 | 0.625 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.3333 | 0.2 | 0.25 | 0.8261 | 0.5758 | 0.6786 | 0.6 | 0.8182 | 0.6923 | 0.6667 | 0.6667 | 0.6667 |
|
| 155 |
+
| 0.201 | 40.0 | 320 | 0.7008 | 1.2940 | 0.7209 | 0.7008 | 0.6824 | 1.0 | 1.0 | 1.0 | 0.8276 | 0.9231 | 0.8727 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 0.0 | 1.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3333 | 0.6667 | 0.4444 | 1.0 | 0.0 | 0.0 | 0.5 | 0.25 | 0.3333 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.6842 | 0.7879 | 0.7324 | 0.6 | 0.5455 | 0.5714 | 0.6667 | 0.6667 | 0.6667 |
|
| 156 |
+
| 0.201 | 41.0 | 328 | 0.7638 | 1.3724 | 0.8040 | 0.7638 | 0.7341 | 1.0 | 1.0 | 1.0 | 0.8 | 0.9231 | 0.8571 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.5 | 0.8333 | 0.625 | 1.0 | 0.0 | 0.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.75 | 0.8571 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.6744 | 0.8788 | 0.7632 | 0.6667 | 0.5455 | 0.6 | 1.0 | 0.6667 | 0.8 |
|
| 157 |
+
| 0.1855 | 42.0 | 336 | 0.7323 | 1.3904 | 0.7521 | 0.7323 | 0.7117 | 1.0 | 1.0 | 1.0 | 0.8846 | 0.8846 | 0.8846 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3636 | 0.6667 | 0.4706 | 1.0 | 0.0 | 0.0 | 0.8 | 0.5 | 0.6154 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.6591 | 0.8788 | 0.7532 | 0.6667 | 0.5455 | 0.6 | 0.6667 | 0.6667 | 0.6667 |
|
| 158 |
+
| 0.1636 | 43.0 | 344 | 0.6929 | 1.4009 | 0.7417 | 0.6929 | 0.6851 | 1.0 | 1.0 | 1.0 | 0.6970 | 0.8846 | 0.7797 | 0.5 | 0.5 | 0.5 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3571 | 0.8333 | 0.5 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.8 | 0.5 | 0.6154 | 1.0 | 0.75 | 0.8571 | 1.0 | 1.0 | 1.0 | 0.7143 | 0.8333 | 0.7692 | 1.0 | 0.0 | 0.0 | 0.2 | 0.2 | 0.2 | 0.7692 | 0.6061 | 0.6780 | 0.7273 | 0.7273 | 0.7273 | 0.6667 | 0.6667 | 0.6667 |
|
| 159 |
+
| 0.1472 | 44.0 | 352 | 0.7165 | 1.4073 | 0.7307 | 0.7165 | 0.7020 | 1.0 | 1.0 | 1.0 | 0.8462 | 0.8462 | 0.8462 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3636 | 0.6667 | 0.4706 | 1.0 | 0.0 | 0.0 | 0.8333 | 0.625 | 0.7143 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.8333 | 0.8333 | 1.0 | 0.5 | 0.6667 | 0.0 | 0.0 | 0.0 | 0.6429 | 0.8182 | 0.72 | 0.625 | 0.4545 | 0.5263 | 0.6667 | 0.6667 | 0.6667 |
|
| 160 |
+
| 0.114 | 45.0 | 360 | 0.7244 | 1.4105 | 0.7553 | 0.7244 | 0.7219 | 1.0 | 1.0 | 1.0 | 0.7059 | 0.9231 | 0.8 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 0.0 | 0.0 | 0.0 | 0.3846 | 0.8333 | 0.5263 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.7143 | 0.625 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.8333 | 0.8333 | 1.0 | 0.5 | 0.6667 | 0.6667 | 0.4 | 0.5 | 0.8077 | 0.6364 | 0.7119 | 0.7 | 0.6364 | 0.6667 | 0.6667 | 0.6667 | 0.6667 |
|
| 161 |
+
| 0.114 | 46.0 | 368 | 0.7402 | 1.3171 | 0.7515 | 0.7402 | 0.7248 | 1.0 | 1.0 | 1.0 | 0.8065 | 0.9615 | 0.8772 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3846 | 0.8333 | 0.5263 | 1.0 | 0.0 | 0.0 | 0.7143 | 0.625 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.5 | 0.6667 | 0.0 | 0.0 | 0.0 | 0.7353 | 0.7576 | 0.7463 | 0.6667 | 0.5455 | 0.6 | 0.6667 | 0.6667 | 0.6667 |
|
| 162 |
+
| 0.0875 | 47.0 | 376 | 0.7402 | 1.3592 | 0.7814 | 0.7402 | 0.7375 | 1.0 | 1.0 | 1.0 | 0.7742 | 0.9231 | 0.8421 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 1.0 | 0.0 | 0.0 | 0.3571 | 0.8333 | 0.5 | 1.0 | 0.0 | 0.0 | 0.7143 | 0.625 | 0.6667 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.5 | 0.6667 | 0.4 | 0.4 | 0.4 | 0.8148 | 0.6667 | 0.7333 | 0.7273 | 0.7273 | 0.7273 | 0.5 | 0.6667 | 0.5714 |
|
| 163 |
+
| 0.0698 | 48.0 | 384 | 0.7244 | 1.4037 | 0.7611 | 0.7244 | 0.7204 | 1.0 | 1.0 | 1.0 | 0.8 | 0.9231 | 0.8571 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 0.0 | 1.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3846 | 0.8333 | 0.5263 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.6667 | 0.5 | 0.5714 | 0.75 | 0.75 | 0.75 | 0.8 | 1.0 | 0.8889 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.5 | 0.6667 | 0.3333 | 0.2 | 0.25 | 0.7667 | 0.6970 | 0.7302 | 0.7 | 0.6364 | 0.6667 | 0.5 | 0.6667 | 0.5714 |
|
| 164 |
+
| 0.0519 | 49.0 | 392 | 0.7244 | 1.5194 | 0.7606 | 0.7244 | 0.7188 | 1.0 | 1.0 | 1.0 | 0.7576 | 0.9615 | 0.8475 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 0.0 | 1.0 | 0.0 | 1.0 | 0.0 | 0.0 | 0.3846 | 0.8333 | 0.5263 | 1.0 | 0.0 | 0.0 | 0.0 | 1.0 | 0.0 | 0.6667 | 0.5 | 0.5714 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.9091 | 1.0 | 0.5 | 0.6667 | 0.3333 | 0.2 | 0.25 | 0.7586 | 0.6667 | 0.7097 | 0.7 | 0.6364 | 0.6667 | 0.6667 | 0.6667 | 0.6667 |
|
| 165 |
+
| 0.0404 | 50.0 | 400 | 0.7087 | 1.5329 | 0.7311 | 0.7087 | 0.7105 | 1.0 | 1.0 | 1.0 | 0.7586 | 0.8462 | 0.8 | 1.0 | 0.5 | 0.6667 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.8571 | 0.75 | 0.8 | 0.0 | 1.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.3846 | 0.8333 | 0.5263 | 0.0 | 0.0 | 0.0 | 0.6667 | 0.5 | 0.5714 | 0.75 | 0.75 | 0.75 | 1.0 | 1.0 | 1.0 | 0.8333 | 0.8333 | 0.8333 | 1.0 | 0.5 | 0.6667 | 0.5 | 0.4 | 0.4444 | 0.7667 | 0.6970 | 0.7302 | 0.6667 | 0.5455 | 0.6 | 0.5 | 0.6667 | 0.5714 |
|
| 166 |
+
|
| 167 |
+
|
| 168 |
+
### Framework versions
|
| 169 |
+
|
| 170 |
+
- Transformers 4.38.2
|
| 171 |
+
- Pytorch 2.2.1+cu121
|
| 172 |
+
- Datasets 2.18.0
|
| 173 |
+
- Tokenizers 0.15.2
|
config.json
ADDED
|
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "nasa-impact/nasa-smd-ibm-st",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"RobertaForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"bos_token_id": 0,
|
| 8 |
+
"classifier_dropout": null,
|
| 9 |
+
"eos_token_id": 2,
|
| 10 |
+
"hidden_act": "gelu",
|
| 11 |
+
"hidden_dropout_prob": 0.1,
|
| 12 |
+
"hidden_size": 768,
|
| 13 |
+
"id2label": {
|
| 14 |
+
"0": "LABEL_0",
|
| 15 |
+
"1": "LABEL_1",
|
| 16 |
+
"2": "LABEL_2",
|
| 17 |
+
"3": "LABEL_3",
|
| 18 |
+
"4": "LABEL_4",
|
| 19 |
+
"5": "LABEL_5",
|
| 20 |
+
"6": "LABEL_6",
|
| 21 |
+
"7": "LABEL_7",
|
| 22 |
+
"8": "LABEL_8",
|
| 23 |
+
"9": "LABEL_9",
|
| 24 |
+
"10": "LABEL_10",
|
| 25 |
+
"11": "LABEL_11",
|
| 26 |
+
"12": "LABEL_12",
|
| 27 |
+
"13": "LABEL_13",
|
| 28 |
+
"14": "LABEL_14",
|
| 29 |
+
"15": "LABEL_15",
|
| 30 |
+
"16": "LABEL_16",
|
| 31 |
+
"17": "LABEL_17",
|
| 32 |
+
"18": "LABEL_18",
|
| 33 |
+
"19": "LABEL_19",
|
| 34 |
+
"20": "LABEL_20",
|
| 35 |
+
"21": "LABEL_21",
|
| 36 |
+
"22": "LABEL_22"
|
| 37 |
+
},
|
| 38 |
+
"initializer_range": 0.02,
|
| 39 |
+
"intermediate_size": 3072,
|
| 40 |
+
"label2id": {
|
| 41 |
+
"LABEL_0": 0,
|
| 42 |
+
"LABEL_1": 1,
|
| 43 |
+
"LABEL_10": 10,
|
| 44 |
+
"LABEL_11": 11,
|
| 45 |
+
"LABEL_12": 12,
|
| 46 |
+
"LABEL_13": 13,
|
| 47 |
+
"LABEL_14": 14,
|
| 48 |
+
"LABEL_15": 15,
|
| 49 |
+
"LABEL_16": 16,
|
| 50 |
+
"LABEL_17": 17,
|
| 51 |
+
"LABEL_18": 18,
|
| 52 |
+
"LABEL_19": 19,
|
| 53 |
+
"LABEL_2": 2,
|
| 54 |
+
"LABEL_20": 20,
|
| 55 |
+
"LABEL_21": 21,
|
| 56 |
+
"LABEL_22": 22,
|
| 57 |
+
"LABEL_3": 3,
|
| 58 |
+
"LABEL_4": 4,
|
| 59 |
+
"LABEL_5": 5,
|
| 60 |
+
"LABEL_6": 6,
|
| 61 |
+
"LABEL_7": 7,
|
| 62 |
+
"LABEL_8": 8,
|
| 63 |
+
"LABEL_9": 9
|
| 64 |
+
},
|
| 65 |
+
"layer_norm_eps": 1e-05,
|
| 66 |
+
"max_position_embeddings": 514,
|
| 67 |
+
"model_type": "roberta",
|
| 68 |
+
"num_attention_heads": 12,
|
| 69 |
+
"num_hidden_layers": 12,
|
| 70 |
+
"pad_token_id": 1,
|
| 71 |
+
"position_embedding_type": "absolute",
|
| 72 |
+
"problem_type": "single_label_classification",
|
| 73 |
+
"torch_dtype": "float32",
|
| 74 |
+
"transformers_version": "4.38.2",
|
| 75 |
+
"type_vocab_size": 1,
|
| 76 |
+
"use_cache": true,
|
| 77 |
+
"vocab_size": 50265
|
| 78 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ca30f0077594be566569fa389ac7ed800940ae40a7b53d5ae99db8692dc979bb
|
| 3 |
+
size 498677420
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f331127a7e9dbbf4206c884e5c479ca2da19ec6757b0978d41d0e3f027115776
|
| 3 |
+
size 4920
|