Tokyosaurus commited on
Commit
e2a89f8
·
verified ·
1 Parent(s): 051f0ce

Model save

Browse files
Files changed (4) hide show
  1. README.md +15 -12
  2. model.safetensors +1 -1
  3. tokenizer.json +2 -2
  4. tokenizer_config.json +3 -44
README.md CHANGED
@@ -19,9 +19,9 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.3824
23
- - Accuracy: 0.9121
24
- - F1: 0.9120
25
 
26
  ## Model description
27
 
@@ -41,25 +41,28 @@ More information needed
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 2e-05
44
- - train_batch_size: 8
45
- - eval_batch_size: 8
46
  - seed: 42
47
- - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
- - num_epochs: 3
 
50
 
51
  ### Training results
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
54
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
55
- | 0.5257 | 1.0 | 102 | 0.3601 | 0.9118 | 0.9174 |
56
- | 0.2449 | 2.0 | 204 | 0.3734 | 0.9314 | 0.9381 |
57
- | 0.2091 | 3.0 | 306 | 0.3486 | 0.9314 | 0.9369 |
 
 
58
 
59
 
60
  ### Framework versions
61
 
62
- - Transformers 4.57.5
63
- - Pytorch 2.9.1+cpu
64
  - Datasets 4.5.0
65
  - Tokenizers 0.22.2
 
19
 
20
  This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.5161
23
+ - Accuracy: 0.8938
24
+ - F1: 0.8910
25
 
26
  ## Model description
27
 
 
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 2e-05
44
+ - train_batch_size: 16
45
+ - eval_batch_size: 16
46
  - seed: 42
47
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
+ - num_epochs: 5
50
+ - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
55
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
56
+ | 0.6300 | 1.0 | 80 | 0.3917 | 0.8516 | 0.8508 |
57
+ | 0.3762 | 2.0 | 160 | 0.4943 | 0.8132 | 0.7866 |
58
+ | 0.3336 | 3.0 | 240 | 0.4476 | 0.8773 | 0.8743 |
59
+ | 0.2274 | 4.0 | 320 | 0.3879 | 0.9048 | 0.9041 |
60
+ | 0.2204 | 5.0 | 400 | 0.5161 | 0.8938 | 0.8910 |
61
 
62
 
63
  ### Framework versions
64
 
65
+ - Transformers 5.0.0
66
+ - Pytorch 2.7.1+cu118
67
  - Datasets 4.5.0
68
  - Tokenizers 0.22.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f32fd86b7c21e2da22fcebbf19def4d6a9aadf4b38c7e013d31700ace7557626
3
  size 1112204984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4459846fc3685bd89c700f2f7979ed515544f8c966e874d7f1cf40aa21cf1215
3
  size 1112204984
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c088c06cf975b7097e469bd69630cdb0d675c6db1ce3af1042b6e19c6d01f22
3
- size 17082999
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5b633524ba90477daaba16ec27580a08a2856ae0ee8c33d9f5f9358378d3b35
3
+ size 16781751
tokenizer_config.json CHANGED
@@ -1,51 +1,10 @@
1
  {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "<s>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "1": {
12
- "content": "<pad>",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "2": {
20
- "content": "</s>",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "3": {
28
- "content": "<unk>",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "250001": {
36
- "content": "<mask>",
37
- "lstrip": true,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
  "bos_token": "<s>",
45
- "clean_up_tokenization_spaces": false,
46
  "cls_token": "<s>",
47
  "eos_token": "</s>",
48
- "extra_special_tokens": {},
49
  "mask_token": "<mask>",
50
  "model_max_length": 512,
51
  "pad_token": "<pad>",
 
1
  {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  "bos_token": "<s>",
 
5
  "cls_token": "<s>",
6
  "eos_token": "</s>",
7
+ "is_local": false,
8
  "mask_token": "<mask>",
9
  "model_max_length": 512,
10
  "pad_token": "<pad>",