End of training
Browse files- README.md +22 -22
- config.json +1 -6
- model.safetensors +2 -2
README.md
CHANGED
|
@@ -14,13 +14,13 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 14 |
|
| 15 |
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
|
| 16 |
It achieves the following results on the evaluation set:
|
| 17 |
-
- Loss:
|
| 18 |
-
- F1 Macro: 0.
|
| 19 |
-
- Precision Macro: 0.
|
| 20 |
-
- Recall Macro: 0.
|
| 21 |
-
- F1 Micro: 0.
|
| 22 |
-
- Precision Micro: 0.
|
| 23 |
-
- Recall Micro: 0.
|
| 24 |
|
| 25 |
## Model description
|
| 26 |
|
|
@@ -52,21 +52,21 @@ The following hyperparameters were used during training:
|
|
| 52 |
|
| 53 |
| Training Loss | Epoch | Step | Validation Loss | F1 Macro | Precision Macro | Recall Macro | F1 Micro | Precision Micro | Recall Micro |
|
| 54 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:---------------:|:------------:|:--------:|:---------------:|:------------:|
|
| 55 |
-
|
|
| 56 |
-
|
|
| 57 |
-
|
|
| 58 |
-
|
|
| 59 |
-
|
|
| 60 |
-
|
|
| 61 |
-
|
|
| 62 |
-
|
|
| 63 |
-
|
|
| 64 |
-
| 0.
|
| 65 |
-
|
|
| 66 |
-
|
|
| 67 |
-
| 0.
|
| 68 |
-
|
|
| 69 |
-
| 0.
|
| 70 |
|
| 71 |
|
| 72 |
### Framework versions
|
|
|
|
| 14 |
|
| 15 |
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
|
| 16 |
It achieves the following results on the evaluation set:
|
| 17 |
+
- Loss: 0.4198
|
| 18 |
+
- F1 Macro: 0.0
|
| 19 |
+
- Precision Macro: 0.0
|
| 20 |
+
- Recall Macro: 0.0
|
| 21 |
+
- F1 Micro: 0.0
|
| 22 |
+
- Precision Micro: 0.0
|
| 23 |
+
- Recall Micro: 0.0
|
| 24 |
|
| 25 |
## Model description
|
| 26 |
|
|
|
|
| 52 |
|
| 53 |
| Training Loss | Epoch | Step | Validation Loss | F1 Macro | Precision Macro | Recall Macro | F1 Micro | Precision Micro | Recall Micro |
|
| 54 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:---------------:|:------------:|:--------:|:---------------:|:------------:|
|
| 55 |
+
| 0.66 | 1.0 | 2 | 0.5815 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 56 |
+
| 0.547 | 2.0 | 4 | 0.5486 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 57 |
+
| 0.5301 | 3.0 | 6 | 0.5299 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 58 |
+
| 0.5108 | 4.0 | 8 | 0.5097 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 59 |
+
| 0.488 | 5.0 | 10 | 0.4892 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 60 |
+
| 0.4685 | 6.0 | 12 | 0.4751 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 61 |
+
| 0.4426 | 7.0 | 14 | 0.4647 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 62 |
+
| 0.4519 | 8.0 | 16 | 0.4542 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 63 |
+
| 0.4134 | 9.0 | 18 | 0.4444 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 64 |
+
| 0.4029 | 10.0 | 20 | 0.4363 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 65 |
+
| 0.3892 | 11.0 | 22 | 0.4297 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 66 |
+
| 0.4437 | 12.0 | 24 | 0.4252 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 67 |
+
| 0.4007 | 13.0 | 26 | 0.4223 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 68 |
+
| 0.4018 | 14.0 | 28 | 0.4205 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 69 |
+
| 0.4107 | 15.0 | 30 | 0.4198 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
|
| 70 |
|
| 71 |
|
| 72 |
### Framework versions
|
config.json
CHANGED
|
@@ -4,12 +4,7 @@
|
|
| 4 |
],
|
| 5 |
"model_type": "bert_model",
|
| 6 |
"num_classes": 4,
|
| 7 |
-
"pos_weight":
|
| 8 |
-
24.0,
|
| 9 |
-
15.666666666666666,
|
| 10 |
-
4.555555555555555,
|
| 11 |
-
6.142857142857143
|
| 12 |
-
],
|
| 13 |
"torch_dtype": "float32",
|
| 14 |
"transformers_version": "4.47.0"
|
| 15 |
}
|
|
|
|
| 4 |
],
|
| 5 |
"model_type": "bert_model",
|
| 6 |
"num_classes": 4,
|
| 7 |
+
"pos_weight": null,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
"torch_dtype": "float32",
|
| 9 |
"transformers_version": "4.47.0"
|
| 10 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:02438d595c6cc2a6c817f9b9f7483ebba9edc253325362fc02c613de353c8e42
|
| 3 |
+
size 437964784
|