Harshil13 commited on
Commit
7f300dd
·
1 Parent(s): 1327cb6

Upload model

Browse files
Files changed (3) hide show
  1. README.md +16 -13
  2. config.json +1 -1
  3. tf_model.h5 +2 -2
README.md CHANGED
@@ -14,13 +14,13 @@ probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - Train Loss: nan
18
- - Train Accuracy: 0.0
19
- - Train Perplexity: 58011.8711
20
- - Validation Loss: 0.2925
21
  - Validation Accuracy: 0.0
22
- - Validation Perplexity: 57475.0
23
- - Epoch: 5
24
 
25
  ## Model description
26
 
@@ -39,19 +39,22 @@ More information needed
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
- - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'WarmUp', 'config': {'initial_learning_rate': 1e-05, 'decay_schedule_fn': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 1e-05, 'decay_steps': 18724, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, '__passive_serialization__': True}, 'warmup_steps': 5000, 'power': 1.0, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
43
  - training_precision: mixed_float16
44
 
45
  ### Training results
46
 
47
  | Train Loss | Train Accuracy | Train Perplexity | Validation Loss | Validation Accuracy | Validation Perplexity | Epoch |
48
  |:----------:|:--------------:|:----------------:|:---------------:|:-------------------:|:---------------------:|:-----:|
49
- | nan | 0.0044 | 133036.0312 | 0.2925 | 0.0 | 57475.0 | 0 |
50
- | nan | 0.0000 | 57758.4336 | 0.2925 | 0.0 | 57475.0 | 1 |
51
- | nan | 0.0000 | 58262.7109 | 0.2925 | 0.0 | 57475.0 | 2 |
52
- | nan | 0.0000 | 57296.8555 | 0.2925 | 0.0 | 57475.0 | 3 |
53
- | nan | 0.0000 | 62598.2734 | 0.2925 | 0.0 | 57475.0 | 4 |
54
- | nan | 0.0 | 58011.8711 | 0.2925 | 0.0 | 57475.0 | 5 |
 
 
 
55
 
56
 
57
  ### Framework versions
 
14
 
15
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 0.3524
18
+ - Train Accuracy: 0.0000
19
+ - Train Perplexity: 18824.3340
20
+ - Validation Loss: 0.3106
21
  - Validation Accuracy: 0.0
22
+ - Validation Perplexity: 39785.5430
23
+ - Epoch: 8
24
 
25
  ## Model description
26
 
 
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'WarmUp', 'config': {'initial_learning_rate': 1e-05, 'decay_schedule_fn': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 1e-05, 'decay_steps': 16381, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, '__passive_serialization__': True}, 'warmup_steps': 1000, 'power': 1.0, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
43
  - training_precision: mixed_float16
44
 
45
  ### Training results
46
 
47
  | Train Loss | Train Accuracy | Train Perplexity | Validation Loss | Validation Accuracy | Validation Perplexity | Epoch |
48
  |:----------:|:--------------:|:----------------:|:---------------:|:-------------------:|:---------------------:|:-----:|
49
+ | 0.6295 | 0.0032 | 100042.4062 | 0.3106 | 0.0 | 39785.5273 | 0 |
50
+ | 0.3528 | 0.0000 | 18560.1328 | 0.3106 | 0.0 | 39785.5391 | 1 |
51
+ | 0.3525 | 0.0000 | 18773.9668 | 0.3106 | 0.0 | 39785.5156 | 2 |
52
+ | 0.3525 | 0.0 | 18342.8223 | 0.3106 | 0.0 | 39785.5078 | 3 |
53
+ | 0.3525 | 0.0000 | 19026.9180 | 0.3106 | 0.0 | 39785.5508 | 4 |
54
+ | 0.3526 | 0.0 | 19108.625 | 0.3106 | 0.0 | 39785.5195 | 5 |
55
+ | 0.3526 | 0.0000 | 19143.7520 | 0.3106 | 0.0 | 39785.5312 | 6 |
56
+ | 0.3525 | 0.0000 | 18503.0938 | 0.3106 | 0.0 | 39785.5195 | 7 |
57
+ | 0.3524 | 0.0000 | 18824.3340 | 0.3106 | 0.0 | 39785.5430 | 8 |
58
 
59
 
60
  ### Framework versions
config.json CHANGED
@@ -34,5 +34,5 @@
34
  },
35
  "transformers_version": "4.26.0",
36
  "use_cache": true,
37
- "vocab_size": 1787
38
  }
 
34
  },
35
  "transformers_version": "4.26.0",
36
  "use_cache": true,
37
+ "vocab_size": 1814
38
  }
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23aff9d575f0e43fc84cf59a7eface16326e0e830187fafe1f919bb8768ef763
3
- size 349035600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc4b029f38898205c326f6444d735540115f93d540fc5018990920f170cb2416
3
+ size 349118544