xb0129 commited on
Commit
31359c0
·
1 Parent(s): 5f0be34

Training in progress epoch 0

Browse files
Files changed (5) hide show
  1. README.md +10 -13
  2. config.json +1 -1
  3. tf_model.h5 +2 -2
  4. tokenizer.json +10 -10
  5. tokenizer_config.json +1 -0
README.md CHANGED
@@ -14,10 +14,10 @@ probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - Train Loss: 0.1696
18
- - Validation Loss: 1.1571
19
- - Train Accuracy: 0.6723
20
- - Epoch: 3
21
 
22
  ## Model description
23
 
@@ -36,22 +36,19 @@ More information needed
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
- - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': False, 'is_legacy_optimizer': False, 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 788, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
40
  - training_precision: float32
41
 
42
  ### Training results
43
 
44
  | Train Loss | Validation Loss | Train Accuracy | Epoch |
45
  |:----------:|:---------------:|:--------------:|:-----:|
46
- | 1.1664 | 1.2814 | 0.5723 | 0 |
47
- | 0.3526 | 1.1862 | 0.6766 | 1 |
48
- | 0.2239 | 1.1143 | 0.6915 | 2 |
49
- | 0.1696 | 1.1571 | 0.6723 | 3 |
50
 
51
 
52
  ### Framework versions
53
 
54
- - Transformers 4.27.4
55
- - TensorFlow 2.11.0
56
- - Datasets 2.11.0
57
- - Tokenizers 0.13.3
 
14
 
15
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 1.1841
18
+ - Validation Loss: 1.4953
19
+ - Train Accuracy: 0.5085
20
+ - Epoch: 0
21
 
22
  ## Model description
23
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
+ - optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 597, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
40
  - training_precision: float32
41
 
42
  ### Training results
43
 
44
  | Train Loss | Validation Loss | Train Accuracy | Epoch |
45
  |:----------:|:---------------:|:--------------:|:-----:|
46
+ | 1.1841 | 1.4953 | 0.5085 | 0 |
 
 
 
47
 
48
 
49
  ### Framework versions
50
 
51
+ - Transformers 4.24.0
52
+ - TensorFlow 2.10.0
53
+ - Datasets 2.10.1
54
+ - Tokenizers 0.11.0
config.json CHANGED
@@ -44,6 +44,6 @@
44
  "seq_classif_dropout": 0.2,
45
  "sinusoidal_pos_embds": false,
46
  "tie_weights_": true,
47
- "transformers_version": "4.27.4",
48
  "vocab_size": 30522
49
  }
 
44
  "seq_classif_dropout": 0.2,
45
  "sinusoidal_pos_embds": false,
46
  "tie_weights_": true,
47
+ "transformers_version": "4.24.0",
48
  "vocab_size": 30522
49
  }
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:415c90e7f9810084581b75e1f2a35825381d272b22c3814f0eaed2a5b220f29c
3
- size 267979456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ad5cc1b6a6047ea60b909abf2c1b173d8bd92e43f4b78ebd22aa4e85847e48e
3
+ size 267982792
tokenizer.json CHANGED
@@ -10,48 +10,48 @@
10
  "added_tokens": [
11
  {
12
  "id": 0,
 
13
  "content": "[PAD]",
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
- "normalized": false,
18
- "special": true
19
  },
20
  {
21
  "id": 100,
 
22
  "content": "[UNK]",
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
- "normalized": false,
27
- "special": true
28
  },
29
  {
30
  "id": 101,
 
31
  "content": "[CLS]",
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
- "normalized": false,
36
- "special": true
37
  },
38
  {
39
  "id": 102,
 
40
  "content": "[SEP]",
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
- "normalized": false,
45
- "special": true
46
  },
47
  {
48
  "id": 103,
 
49
  "content": "[MASK]",
50
  "single_word": false,
51
  "lstrip": false,
52
  "rstrip": false,
53
- "normalized": false,
54
- "special": true
55
  }
56
  ],
57
  "normalizer": {
 
10
  "added_tokens": [
11
  {
12
  "id": 0,
13
+ "special": true,
14
  "content": "[PAD]",
15
  "single_word": false,
16
  "lstrip": false,
17
  "rstrip": false,
18
+ "normalized": false
 
19
  },
20
  {
21
  "id": 100,
22
+ "special": true,
23
  "content": "[UNK]",
24
  "single_word": false,
25
  "lstrip": false,
26
  "rstrip": false,
27
+ "normalized": false
 
28
  },
29
  {
30
  "id": 101,
31
+ "special": true,
32
  "content": "[CLS]",
33
  "single_word": false,
34
  "lstrip": false,
35
  "rstrip": false,
36
+ "normalized": false
 
37
  },
38
  {
39
  "id": 102,
40
+ "special": true,
41
  "content": "[SEP]",
42
  "single_word": false,
43
  "lstrip": false,
44
  "rstrip": false,
45
+ "normalized": false
 
46
  },
47
  {
48
  "id": 103,
49
+ "special": true,
50
  "content": "[MASK]",
51
  "single_word": false,
52
  "lstrip": false,
53
  "rstrip": false,
54
+ "normalized": false
 
55
  }
56
  ],
57
  "normalizer": {
tokenizer_config.json CHANGED
@@ -3,6 +3,7 @@
3
  "do_lower_case": true,
4
  "mask_token": "[MASK]",
5
  "model_max_length": 512,
 
6
  "pad_token": "[PAD]",
7
  "sep_token": "[SEP]",
8
  "special_tokens_map_file": null,
 
3
  "do_lower_case": true,
4
  "mask_token": "[MASK]",
5
  "model_max_length": 512,
6
+ "name_or_path": "distilbert-base-uncased",
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
9
  "special_tokens_map_file": null,