trungpq commited on
Commit
2aa4437
·
verified ·
1 Parent(s): ede82c6

End of training

Browse files
Files changed (4) hide show
  1. README.md +34 -31
  2. config.json +1 -1
  3. model.safetensors +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -16,12 +16,12 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 2.7387
20
- - Accuracy: 0.6195
21
- - F1 Macro: 0.6660
22
- - Precision Macro: 0.6882
23
- - Recall Macro: 0.6608
24
- - Total Tf: [254, 156, 1074, 156]
25
 
26
  ## Model description
27
 
@@ -46,39 +46,42 @@ The following hyperparameters were used during training:
46
  - seed: 42
47
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
- - lr_scheduler_warmup_steps: 58
50
  - num_epochs: 25
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Precision Macro | Recall Macro | Total Tf |
55
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:---------------:|:------------:|:---------------------:|
56
- | 1.1297 | 1.0 | 59 | 1.1018 | 0.4707 | 0.4559 | 0.6732 | 0.5156 | [193, 217, 1013, 217] |
57
- | 0.9888 | 2.0 | 118 | 1.0099 | 0.5829 | 0.6305 | 0.6331 | 0.6344 | [239, 171, 1059, 171] |
58
- | 0.6799 | 3.0 | 177 | 1.0688 | 0.5683 | 0.6210 | 0.6181 | 0.6293 | [233, 177, 1053, 177] |
59
- | 0.5096 | 4.0 | 236 | 1.2211 | 0.5707 | 0.6199 | 0.6183 | 0.6332 | [234, 176, 1054, 176] |
60
- | 0.3765 | 5.0 | 295 | 1.2180 | 0.6146 | 0.6566 | 0.6728 | 0.6514 | [252, 158, 1072, 158] |
61
- | 0.2702 | 6.0 | 354 | 1.4150 | 0.6146 | 0.6626 | 0.6780 | 0.6591 | [252, 158, 1072, 158] |
62
- | 0.1631 | 7.0 | 413 | 1.5638 | 0.6171 | 0.6615 | 0.6812 | 0.6586 | [253, 157, 1073, 157] |
63
- | 0.1456 | 8.0 | 472 | 1.7113 | 0.6024 | 0.6507 | 0.6731 | 0.6472 | [247, 163, 1067, 163] |
64
- | 0.1162 | 9.0 | 531 | 1.7476 | 0.6122 | 0.6604 | 0.6713 | 0.6563 | [251, 159, 1071, 159] |
65
- | 0.0759 | 10.0 | 590 | 1.9517 | 0.6073 | 0.6547 | 0.6710 | 0.6521 | [249, 161, 1069, 161] |
66
- | 0.0593 | 11.0 | 649 | 2.0548 | 0.6195 | 0.6678 | 0.6777 | 0.6659 | [254, 156, 1074, 156] |
67
- | 0.0511 | 12.0 | 708 | 2.1851 | 0.6049 | 0.6525 | 0.6889 | 0.6498 | [248, 162, 1068, 162] |
68
- | 0.0483 | 13.0 | 767 | 2.2704 | 0.6098 | 0.6582 | 0.6729 | 0.6560 | [250, 160, 1070, 160] |
69
- | 0.0391 | 14.0 | 826 | 2.3890 | 0.6146 | 0.6607 | 0.6867 | 0.6571 | [252, 158, 1072, 158] |
70
- | 0.0231 | 15.0 | 885 | 2.4750 | 0.6122 | 0.6596 | 0.6908 | 0.6545 | [251, 159, 1071, 159] |
71
- | 0.0104 | 16.0 | 944 | 2.4944 | 0.6293 | 0.6743 | 0.6953 | 0.6687 | [258, 152, 1078, 152] |
72
- | 0.019 | 17.0 | 1003 | 2.6014 | 0.6171 | 0.6649 | 0.6851 | 0.6615 | [253, 157, 1073, 157] |
73
- | 0.0049 | 18.0 | 1062 | 2.5812 | 0.6293 | 0.6738 | 0.6884 | 0.6688 | [258, 152, 1078, 152] |
74
- | 0.0117 | 19.0 | 1121 | 2.6739 | 0.6220 | 0.6681 | 0.6955 | 0.6617 | [255, 155, 1075, 155] |
75
- | 0.0155 | 20.0 | 1180 | 2.6334 | 0.6195 | 0.6636 | 0.6907 | 0.6566 | [254, 156, 1074, 156] |
76
- | 0.0241 | 21.0 | 1239 | 2.7387 | 0.6195 | 0.6660 | 0.6882 | 0.6608 | [254, 156, 1074, 156] |
 
 
 
 
77
 
78
 
79
  ### Framework versions
80
 
81
- - Transformers 4.47.0
82
- - Pytorch 2.5.1+cu121
83
- - Datasets 3.2.0
84
  - Tokenizers 0.21.0
 
16
 
17
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 2.0451
20
+ - Accuracy: 0.6463
21
+ - F1 Macro: 0.6925
22
+ - Precision Macro: 0.6941
23
+ - Recall Macro: 0.6925
24
+ - Total Tf: [265, 145, 1085, 145]
25
 
26
  ## Model description
27
 
 
46
  - seed: 42
47
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
+ - lr_scheduler_warmup_steps: 90
50
  - num_epochs: 25
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Precision Macro | Recall Macro | Total Tf |
55
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:---------------:|:------------:|:---------------------:|
56
+ | 1.0951 | 1.0 | 91 | 1.1126 | 0.4366 | 0.4534 | 0.6658 | 0.5044 | [179, 231, 999, 231] |
57
+ | 0.9371 | 2.0 | 182 | 0.9998 | 0.5756 | 0.5422 | 0.5071 | 0.5904 | [236, 174, 1056, 174] |
58
+ | 0.7416 | 3.0 | 273 | 1.0900 | 0.5488 | 0.5805 | 0.5892 | 0.6078 | [225, 185, 1045, 185] |
59
+ | 0.6429 | 4.0 | 364 | 1.1685 | 0.5634 | 0.5517 | 0.4949 | 0.6483 | [231, 179, 1051, 179] |
60
+ | 0.5992 | 5.0 | 455 | 1.0657 | 0.6366 | 0.6595 | 0.6879 | 0.6604 | [261, 149, 1081, 149] |
61
+ | 0.5504 | 6.0 | 546 | 1.1434 | 0.6463 | 0.6839 | 0.7002 | 0.6790 | [265, 145, 1085, 145] |
62
+ | 0.4743 | 7.0 | 637 | 1.1307 | 0.6415 | 0.6852 | 0.6817 | 0.6951 | [263, 147, 1083, 147] |
63
+ | 0.3508 | 8.0 | 728 | 1.2377 | 0.6244 | 0.6679 | 0.6637 | 0.6810 | [256, 154, 1076, 154] |
64
+ | 0.3027 | 9.0 | 819 | 1.3096 | 0.6366 | 0.6821 | 0.6824 | 0.6818 | [261, 149, 1081, 149] |
65
+ | 0.2456 | 10.0 | 910 | 1.3739 | 0.6537 | 0.6986 | 0.6975 | 0.7006 | [268, 142, 1088, 142] |
66
+ | 0.2111 | 11.0 | 1001 | 1.4215 | 0.6341 | 0.6809 | 0.6801 | 0.6821 | [260, 150, 1080, 150] |
67
+ | 0.1452 | 12.0 | 1092 | 1.5324 | 0.6366 | 0.6822 | 0.6832 | 0.6814 | [261, 149, 1081, 149] |
68
+ | 0.1426 | 13.0 | 1183 | 1.6104 | 0.6415 | 0.6871 | 0.6918 | 0.6845 | [263, 147, 1083, 147] |
69
+ | 0.142 | 14.0 | 1274 | 1.6417 | 0.6390 | 0.6852 | 0.6846 | 0.6867 | [262, 148, 1082, 148] |
70
+ | 0.1004 | 15.0 | 1365 | 1.7112 | 0.6439 | 0.6904 | 0.6945 | 0.6896 | [264, 146, 1084, 146] |
71
+ | 0.1292 | 16.0 | 1456 | 1.7041 | 0.6463 | 0.6932 | 0.6974 | 0.6921 | [265, 145, 1085, 145] |
72
+ | 0.0998 | 17.0 | 1547 | 1.7698 | 0.6512 | 0.6956 | 0.6951 | 0.6964 | [267, 143, 1087, 143] |
73
+ | 0.073 | 18.0 | 1638 | 1.8860 | 0.6488 | 0.6948 | 0.7013 | 0.6919 | [266, 144, 1086, 144] |
74
+ | 0.0736 | 19.0 | 1729 | 1.9039 | 0.6390 | 0.6859 | 0.6875 | 0.6854 | [262, 148, 1082, 148] |
75
+ | 0.0548 | 20.0 | 1820 | 2.0032 | 0.6366 | 0.6844 | 0.6855 | 0.6864 | [261, 149, 1081, 149] |
76
+ | 0.0554 | 21.0 | 1911 | 2.0158 | 0.6317 | 0.6804 | 0.6843 | 0.6807 | [259, 151, 1079, 151] |
77
+ | 0.0583 | 22.0 | 2002 | 2.0387 | 0.6439 | 0.6908 | 0.6949 | 0.6903 | [264, 146, 1084, 146] |
78
+ | 0.0427 | 23.0 | 2093 | 2.0383 | 0.6512 | 0.6965 | 0.6992 | 0.6962 | [267, 143, 1087, 143] |
79
+ | 0.0453 | 24.0 | 2184 | 2.0345 | 0.6463 | 0.6925 | 0.6941 | 0.6925 | [265, 145, 1085, 145] |
80
+ | 0.0466 | 25.0 | 2275 | 2.0451 | 0.6463 | 0.6925 | 0.6941 | 0.6925 | [265, 145, 1085, 145] |
81
 
82
 
83
  ### Framework versions
84
 
85
+ - Transformers 4.48.3
86
+ - Pytorch 2.1.0+cu118
 
87
  - Tokenizers 0.21.0
config.json CHANGED
@@ -8,5 +8,5 @@
8
  "model_type": "bert_with_absa",
9
  "num_classes": 3,
10
  "torch_dtype": "float32",
11
- "transformers_version": "4.47.0"
12
  }
 
8
  "model_type": "bert_with_absa",
9
  "num_classes": 3,
10
  "torch_dtype": "float32",
11
+ "transformers_version": "4.48.3"
12
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:484af7886f8fa07cf8c198b5e6069d25d2de73f27df3f21a6014258ead963112
3
- size 875933728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d9d31eb1654ec0b2ae94b732491ef97aff1ed535dc74ac6029e5d5a08de6adc
3
+ size 875933980
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7270ee1260397fb9d635e797fa84517f04958df7cfd38871f98980047d75ca42
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:785c2afb2cf11154f2c1ed6c58607806e88e8f047d32f2e48c607d809d56abec
3
  size 5368