santoshmds21 commited on
Commit
af8ac86
·
verified ·
1 Parent(s): e71de13

End of training

Browse files
Files changed (5) hide show
  1. README.md +18 -18
  2. config.json +8 -3
  3. model.safetensors +2 -2
  4. tokenizer_config.json +2 -44
  5. training_args.bin +2 -2
README.md CHANGED
@@ -18,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.7047
22
- - Accuracy: 0.491
23
- - Auc: 0.75
24
 
25
  ## Model description
26
 
@@ -43,7 +43,7 @@ The following hyperparameters were used during training:
43
  - train_batch_size: 8
44
  - eval_batch_size: 8
45
  - seed: 42
46
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - num_epochs: 10
49
 
@@ -51,21 +51,21 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Auc |
53
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:-----:|
54
- | 0.7135 | 1.0 | 263 | 0.6957 | 0.509 | 0.692 |
55
- | 0.7053 | 2.0 | 526 | 0.7073 | 0.491 | 0.274 |
56
- | 0.7033 | 3.0 | 789 | 0.7039 | 0.509 | 0.701 |
57
- | 0.7025 | 4.0 | 1052 | 0.6955 | 0.491 | 0.471 |
58
- | 0.6995 | 5.0 | 1315 | 0.7008 | 0.491 | 0.533 |
59
- | 0.6993 | 6.0 | 1578 | 0.6982 | 0.491 | 0.708 |
60
- | 0.696 | 7.0 | 1841 | 0.6993 | 0.491 | 0.654 |
61
- | 0.6939 | 8.0 | 2104 | 0.6954 | 0.491 | 0.705 |
62
- | 0.6907 | 9.0 | 2367 | 0.6994 | 0.491 | 0.673 |
63
- | 0.6946 | 10.0 | 2630 | 0.7047 | 0.491 | 0.75 |
64
 
65
 
66
  ### Framework versions
67
 
68
- - Transformers 4.52.4
69
- - Pytorch 2.6.0+cu124
70
- - Datasets 3.6.0
71
- - Tokenizers 0.21.1
 
18
 
19
  This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.2897
22
+ - Accuracy: 0.864
23
+ - Auc: 0.951
24
 
25
  ## Model description
26
 
 
43
  - train_batch_size: 8
44
  - eval_batch_size: 8
45
  - seed: 42
46
+ - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - num_epochs: 10
49
 
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Auc |
53
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:-----:|
54
+ | 0.5042 | 1.0 | 263 | 0.3865 | 0.813 | 0.912 |
55
+ | 0.4105 | 2.0 | 526 | 0.3380 | 0.847 | 0.931 |
56
+ | 0.3583 | 3.0 | 789 | 0.3148 | 0.856 | 0.939 |
57
+ | 0.3553 | 4.0 | 1052 | 0.3454 | 0.851 | 0.945 |
58
+ | 0.3511 | 5.0 | 1315 | 0.3450 | 0.86 | 0.948 |
59
+ | 0.3477 | 6.0 | 1578 | 0.2906 | 0.871 | 0.95 |
60
+ | 0.3346 | 7.0 | 1841 | 0.2879 | 0.876 | 0.95 |
61
+ | 0.3096 | 8.0 | 2104 | 0.2892 | 0.869 | 0.95 |
62
+ | 0.3153 | 9.0 | 2367 | 0.2841 | 0.88 | 0.951 |
63
+ | 0.3140 | 10.0 | 2630 | 0.2897 | 0.864 | 0.951 |
64
 
65
 
66
  ### Framework versions
67
 
68
+ - Transformers 5.0.0
69
+ - Pytorch 2.9.0+cu128
70
+ - Datasets 4.0.0
71
+ - Tokenizers 0.22.2
config.json CHANGED
@@ -1,9 +1,13 @@
1
  {
 
2
  "architectures": [
3
  "BertForSequenceClassification"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
 
6
  "classifier_dropout": null,
 
 
7
  "gradient_checkpointing": false,
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
@@ -14,6 +18,7 @@
14
  },
15
  "initializer_range": 0.02,
16
  "intermediate_size": 3072,
 
17
  "label2id": {
18
  "Not Safe": 1,
19
  "Safe": 0
@@ -26,9 +31,9 @@
26
  "pad_token_id": 0,
27
  "position_embedding_type": "absolute",
28
  "problem_type": "single_label_classification",
29
- "torch_dtype": "float32",
30
- "transformers_version": "4.52.4",
31
  "type_vocab_size": 2,
32
- "use_cache": true,
33
  "vocab_size": 30522
34
  }
 
1
  {
2
+ "add_cross_attention": false,
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
  "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": null,
11
  "gradient_checkpointing": false,
12
  "hidden_act": "gelu",
13
  "hidden_dropout_prob": 0.1,
 
18
  },
19
  "initializer_range": 0.02,
20
  "intermediate_size": 3072,
21
+ "is_decoder": false,
22
  "label2id": {
23
  "Not Safe": 1,
24
  "Safe": 0
 
31
  "pad_token_id": 0,
32
  "position_embedding_type": "absolute",
33
  "problem_type": "single_label_classification",
34
+ "tie_word_embeddings": true,
35
+ "transformers_version": "5.0.0",
36
  "type_vocab_size": 2,
37
+ "use_cache": false,
38
  "vocab_size": 30522
39
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6df64b9931b12c83211a9ba030e7d537ddd03b464a1d8f0d317147b5f34ea82
3
- size 437958648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5420ca952cc43dc0d622129400eb0245c8e79708adefac6859ec5dcd5e4e4383
3
+ size 437958624
tokenizer_config.json CHANGED
@@ -1,50 +1,8 @@
1
  {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "100": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "101": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "102": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "103": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
- "clean_up_tokenization_spaces": false,
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
- "extra_special_tokens": {},
48
  "mask_token": "[MASK]",
49
  "model_max_length": 512,
50
  "pad_token": "[PAD]",
 
1
  {
2
+ "backend": "tokenizers",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "cls_token": "[CLS]",
4
  "do_lower_case": true,
5
+ "is_local": false,
6
  "mask_token": "[MASK]",
7
  "model_max_length": 512,
8
  "pad_token": "[PAD]",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09a6e9bc44b14e5c67ec1b7b660450f8bb920def366f7a2218c5ae1db00e4764
3
- size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f59c3ead0ade00341beb61ed9e00321ae520b2f6f46fcfb3ac32e5e0f011b17
3
+ size 5201