trungpq commited on
Commit
6435973
·
verified ·
1 Parent(s): aa22f21

End of training

Browse files
Files changed (4) hide show
  1. README.md +82 -0
  2. config.json +12 -0
  3. model.safetensors +3 -0
  4. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - generated_from_trainer
5
+ metrics:
6
+ - accuracy
7
+ model-index:
8
+ - name: rlcc-palate-upsample_replacement-absa-max
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # rlcc-palate-upsample_replacement-absa-max
16
+
17
+ This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 1.9359
20
+ - Accuracy: 0.8
21
+ - F1 Macro: 0.5653
22
+ - Precision Macro: 0.5691
23
+ - Recall Macro: 0.6086
24
+ - Total Tf: [328, 82, 1148, 82]
25
+
26
+ ## Model description
27
+
28
+ More information needed
29
+
30
+ ## Intended uses & limitations
31
+
32
+ More information needed
33
+
34
+ ## Training and evaluation data
35
+
36
+ More information needed
37
+
38
+ ## Training procedure
39
+
40
+ ### Training hyperparameters
41
+
42
+ The following hyperparameters were used during training:
43
+ - learning_rate: 2e-05
44
+ - train_batch_size: 64
45
+ - eval_batch_size: 64
46
+ - seed: 42
47
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
+ - lr_scheduler_type: linear
49
+ - lr_scheduler_warmup_steps: 36
50
+ - num_epochs: 25
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Precision Macro | Recall Macro | Total Tf |
55
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:---------------:|:------------:|:-------------------:|
56
+ | 1.1087 | 1.0 | 37 | 1.0814 | 0.8 | 0.4854 | 0.4886 | 0.4866 | [328, 82, 1148, 82] |
57
+ | 1.0809 | 2.0 | 74 | 1.0422 | 0.8366 | 0.4819 | 0.4582 | 0.5129 | [343, 67, 1163, 67] |
58
+ | 0.9443 | 3.0 | 111 | 1.0914 | 0.8195 | 0.5403 | 0.5491 | 0.5776 | [336, 74, 1156, 74] |
59
+ | 0.7914 | 4.0 | 148 | 1.1077 | 0.8171 | 0.5736 | 0.5725 | 0.6223 | [335, 75, 1155, 75] |
60
+ | 0.68 | 5.0 | 185 | 1.2576 | 0.7585 | 0.4613 | 0.4035 | 0.6144 | [311, 99, 1131, 99] |
61
+ | 0.5892 | 6.0 | 222 | 1.2538 | 0.8049 | 0.5494 | 0.5447 | 0.5680 | [330, 80, 1150, 80] |
62
+ | 0.564 | 7.0 | 259 | 1.2608 | 0.8098 | 0.5699 | 0.5680 | 0.5977 | [332, 78, 1152, 78] |
63
+ | 0.5395 | 8.0 | 296 | 1.3882 | 0.7585 | 0.4659 | 0.4086 | 0.6166 | [311, 99, 1131, 99] |
64
+ | 0.5745 | 9.0 | 333 | 1.2718 | 0.8098 | 0.5878 | 0.5968 | 0.6340 | [332, 78, 1152, 78] |
65
+ | 0.5492 | 10.0 | 370 | 1.3772 | 0.8122 | 0.5641 | 0.5579 | 0.5838 | [333, 77, 1153, 77] |
66
+ | 0.4613 | 11.0 | 407 | 1.5313 | 0.8293 | 0.5471 | 0.5552 | 0.5915 | [340, 70, 1160, 70] |
67
+ | 0.5209 | 12.0 | 444 | 1.5491 | 0.8244 | 0.5125 | 0.4843 | 0.5547 | [338, 72, 1158, 72] |
68
+ | 0.4865 | 13.0 | 481 | 1.6926 | 0.8073 | 0.5809 | 0.5987 | 0.6386 | [331, 79, 1151, 79] |
69
+ | 0.4758 | 14.0 | 518 | 1.6802 | 0.8195 | 0.6041 | 0.6071 | 0.6511 | [336, 74, 1156, 74] |
70
+ | 0.3792 | 15.0 | 555 | 1.7840 | 0.8098 | 0.5873 | 0.6012 | 0.6491 | [332, 78, 1152, 78] |
71
+ | 0.3555 | 16.0 | 592 | 1.8224 | 0.8122 | 0.5915 | 0.5943 | 0.6423 | [333, 77, 1153, 77] |
72
+ | 0.2957 | 17.0 | 629 | 1.8715 | 0.8098 | 0.5790 | 0.5777 | 0.6106 | [332, 78, 1152, 78] |
73
+ | 0.2693 | 18.0 | 666 | 1.9544 | 0.8 | 0.5683 | 0.5704 | 0.6157 | [328, 82, 1148, 82] |
74
+ | 0.248 | 19.0 | 703 | 1.9359 | 0.8 | 0.5653 | 0.5691 | 0.6086 | [328, 82, 1148, 82] |
75
+
76
+
77
+ ### Framework versions
78
+
79
+ - Transformers 4.47.0
80
+ - Pytorch 2.5.1+cu121
81
+ - Datasets 3.2.0
82
+ - Tokenizers 0.21.0
config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "absa_method": "max",
3
+ "architectures": [
4
+ "BERTModel"
5
+ ],
6
+ "class_weight": null,
7
+ "class_weights": null,
8
+ "model_type": "bert_with_absa",
9
+ "num_classes": 3,
10
+ "torch_dtype": "float32",
11
+ "transformers_version": "4.47.0"
12
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47ee8145ab2f5d8a6d88926149f6f6f7fc5a667a3dc629d6319a68b86ad86672
3
+ size 875933980
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6be25c20ab3c0faf6b87bb8d0b4de1c8e999344571c7b1d703bc8e2ce16aa9e6
3
+ size 5368