cedricbonhomme commited on
Commit
36d9ef6
·
verified ·
1 Parent(s): 430ec8b

End of training

Browse files
Files changed (4) hide show
  1. README.md +75 -0
  2. model.safetensors +1 -1
  3. tokenizer.json +0 -0
  4. tokenizer_config.json +14 -0
README.md ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: hfl/chinese-macbert-base
5
+ tags:
6
+ - generated_from_trainer
7
+ metrics:
8
+ - accuracy
9
+ model-index:
10
+ - name: vulnerability-severity-classification-chinese-macbert-base-test
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # vulnerability-severity-classification-chinese-macbert-base-test
18
+
19
+ This model is a fine-tuned version of [hfl/chinese-macbert-base](https://huggingface.co/hfl/chinese-macbert-base) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.8342
22
+ - Accuracy: 0.7323
23
+ - F1 Macro: 0.6756
24
+ - Low Precision: 0.4336
25
+ - Low Recall: 0.6084
26
+ - Low F1: 0.5064
27
+ - Medium Precision: 0.8150
28
+ - Medium Recall: 0.7024
29
+ - Medium F1: 0.7545
30
+ - High Precision: 0.7263
31
+ - High Recall: 0.8099
32
+ - High F1: 0.7658
33
+
34
+ ## Model description
35
+
36
+ More information needed
37
+
38
+ ## Intended uses & limitations
39
+
40
+ More information needed
41
+
42
+ ## Training and evaluation data
43
+
44
+ More information needed
45
+
46
+ ## Training procedure
47
+
48
+ ### Training hyperparameters
49
+
50
+ The following hyperparameters were used during training:
51
+ - learning_rate: 3e-05
52
+ - train_batch_size: 64
53
+ - eval_batch_size: 64
54
+ - seed: 42
55
+ - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
56
+ - lr_scheduler_type: linear
57
+ - num_epochs: 5
58
+
59
+ ### Training results
60
+
61
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Low Precision | Low Recall | Low F1 | Medium Precision | Medium Recall | Medium F1 | High Precision | High Recall | High F1 |
62
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:-------------:|:----------:|:------:|:----------------:|:-------------:|:---------:|:--------------:|:-----------:|:-------:|
63
+ | 0.7148 | 1.0 | 1590 | 0.7313 | 0.6901 | 0.6344 | 0.3572 | 0.6132 | 0.4514 | 0.7990 | 0.6245 | 0.7010 | 0.6984 | 0.8115 | 0.7507 |
64
+ | 0.7214 | 2.0 | 3180 | 0.6823 | 0.6980 | 0.6448 | 0.3467 | 0.6795 | 0.4591 | 0.8066 | 0.6398 | 0.7136 | 0.7323 | 0.7934 | 0.7616 |
65
+ | 0.4822 | 3.0 | 4770 | 0.6937 | 0.6999 | 0.6464 | 0.3403 | 0.6934 | 0.4566 | 0.8303 | 0.6258 | 0.7137 | 0.7260 | 0.8171 | 0.7688 |
66
+ | 0.5379 | 4.0 | 6360 | 0.7548 | 0.7210 | 0.6653 | 0.3980 | 0.6354 | 0.4894 | 0.8284 | 0.6635 | 0.7369 | 0.7161 | 0.8320 | 0.7697 |
67
+ | 0.3922 | 5.0 | 7950 | 0.8342 | 0.7323 | 0.6756 | 0.4336 | 0.6084 | 0.5064 | 0.8150 | 0.7024 | 0.7545 | 0.7263 | 0.8099 | 0.7658 |
68
+
69
+
70
+ ### Framework versions
71
+
72
+ - Transformers 5.4.0
73
+ - Pytorch 2.11.0+cu130
74
+ - Datasets 4.8.4
75
+ - Tokenizers 0.22.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25f6e69009eadbb31e93040e3f541933333365b0a994a48c68d809eded68e2dd
3
  size 409103292
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2a64ecdcc50f18808948fe678969167b4172c3c2b0e7c7362e0f45ea0e0e564
3
  size 409103292
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }