trungpq commited on
Commit
df46e30
·
verified ·
1 Parent(s): 8f92597

End of training

Browse files
Files changed (4) hide show
  1. README.md +74 -0
  2. config.json +12 -0
  3. model.safetensors +3 -0
  4. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - generated_from_trainer
5
+ metrics:
6
+ - accuracy
7
+ model-index:
8
+ - name: rlcc-palate-upsample_replacement-absa-None
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # rlcc-palate-upsample_replacement-absa-None
16
+
17
+ This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 1.8831
20
+ - Accuracy: 0.8049
21
+ - F1 Macro: 0.5751
22
+ - Precision Macro: 0.5829
23
+ - Recall Macro: 0.6074
24
+ - Total Tf: [330, 80, 1150, 80]
25
+
26
+ ## Model description
27
+
28
+ More information needed
29
+
30
+ ## Intended uses & limitations
31
+
32
+ More information needed
33
+
34
+ ## Training and evaluation data
35
+
36
+ More information needed
37
+
38
+ ## Training procedure
39
+
40
+ ### Training hyperparameters
41
+
42
+ The following hyperparameters were used during training:
43
+ - learning_rate: 2e-05
44
+ - train_batch_size: 64
45
+ - eval_batch_size: 64
46
+ - seed: 42
47
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
+ - lr_scheduler_type: linear
49
+ - lr_scheduler_warmup_steps: 36
50
+ - num_epochs: 25
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Precision Macro | Recall Macro | Total Tf |
55
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:---------------:|:------------:|:-------------------:|
56
+ | 1.1239 | 1.0 | 37 | 1.0929 | 0.7854 | 0.4581 | 0.4644 | 0.4843 | [322, 88, 1142, 88] |
57
+ | 1.0966 | 2.0 | 74 | 1.0095 | 0.8049 | 0.4341 | 0.4847 | 0.4513 | [330, 80, 1150, 80] |
58
+ | 0.9736 | 3.0 | 111 | 1.0467 | 0.8098 | 0.5311 | 0.5396 | 0.5271 | [332, 78, 1152, 78] |
59
+ | 0.7549 | 4.0 | 148 | 1.0651 | 0.7976 | 0.5381 | 0.5417 | 0.5606 | [327, 83, 1147, 83] |
60
+ | 0.6054 | 5.0 | 185 | 1.1613 | 0.8122 | 0.5799 | 0.5767 | 0.6065 | [333, 77, 1153, 77] |
61
+ | 0.4563 | 6.0 | 222 | 1.3234 | 0.8146 | 0.5998 | 0.6069 | 0.6501 | [334, 76, 1154, 76] |
62
+ | 0.3093 | 7.0 | 259 | 1.3972 | 0.8146 | 0.5904 | 0.5941 | 0.6094 | [334, 76, 1154, 76] |
63
+ | 0.1836 | 8.0 | 296 | 1.7261 | 0.8049 | 0.5787 | 0.5880 | 0.6291 | [330, 80, 1150, 80] |
64
+ | 0.1617 | 9.0 | 333 | 1.7743 | 0.8049 | 0.5791 | 0.5951 | 0.6291 | [330, 80, 1150, 80] |
65
+ | 0.1209 | 10.0 | 370 | 1.9516 | 0.8073 | 0.5921 | 0.6081 | 0.6603 | [331, 79, 1151, 79] |
66
+ | 0.0736 | 11.0 | 407 | 1.8831 | 0.8049 | 0.5751 | 0.5829 | 0.6074 | [330, 80, 1150, 80] |
67
+
68
+
69
+ ### Framework versions
70
+
71
+ - Transformers 4.47.0
72
+ - Pytorch 2.5.1+cu121
73
+ - Datasets 3.2.0
74
+ - Tokenizers 0.21.0
config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "absa_method": null,
3
+ "architectures": [
4
+ "BERTModel"
5
+ ],
6
+ "class_weight": null,
7
+ "class_weights": null,
8
+ "model_type": "bert_with_absa",
9
+ "num_classes": 3,
10
+ "torch_dtype": "float32",
11
+ "transformers_version": "4.47.0"
12
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e74bef5cb991e99062428bbbb836e54b9dc3db35ec6574c920e514296c76ff13
3
+ size 437963716
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b45f7722b9f367020c6f43fcfc7558dce3847d12bb4a2fde6fd2335663cdc61
3
+ size 5368