smiled0g commited on
Commit
fb41304
·
verified ·
1 Parent(s): 947a88b

End of training

Browse files
Files changed (4) hide show
  1. README.md +4 -9
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. training_args.bin +2 -2
README.md CHANGED
@@ -10,7 +10,6 @@ model-index:
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
  should probably proofread and complete it, then remove this comment. -->
12
 
13
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/smiled0g/preflop/runs/aue4q75g)
14
  # preflop
15
 
16
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
@@ -33,21 +32,17 @@ More information needed
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 1e-05
36
- - train_batch_size: 128
37
  - eval_batch_size: 8
38
  - seed: 42
39
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
- - num_epochs: 3
42
  - mixed_precision_training: Native AMP
43
 
44
- ### Training results
45
-
46
-
47
-
48
  ### Framework versions
49
 
50
- - Transformers 4.45.2
51
  - Pytorch 2.1.1+cu121
52
  - Datasets 3.0.2
53
  - Tokenizers 0.20.1
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
  should probably proofread and complete it, then remove this comment. -->
12
 
 
13
  # preflop
14
 
15
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
 
32
 
33
  The following hyperparameters were used during training:
34
  - learning_rate: 1e-05
35
+ - train_batch_size: 100
36
  - eval_batch_size: 8
37
  - seed: 42
38
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
39
  - lr_scheduler_type: linear
40
+ - num_epochs: 2
41
  - mixed_precision_training: Native AMP
42
 
 
 
 
 
43
  ### Framework versions
44
 
45
+ - Transformers 4.46.0
46
  - Pytorch 2.1.1+cu121
47
  - Datasets 3.0.2
48
  - Tokenizers 0.20.1
config.json CHANGED
@@ -38,7 +38,7 @@
38
  "rope_theta": 10000.0,
39
  "tie_word_embeddings": false,
40
  "torch_dtype": "float32",
41
- "transformers_version": "4.45.2",
42
  "use_cache": true,
43
  "vocab_size": 53
44
  }
 
38
  "rope_theta": 10000.0,
39
  "tie_word_embeddings": false,
40
  "torch_dtype": "float32",
41
+ "transformers_version": "4.46.0",
42
  "use_cache": true,
43
  "vocab_size": 53
44
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4251b110111ca00d62c5078597269e737702ad49f6e19f7613700852ab4048cb
3
  size 1074131104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a2a6beab37ccf7a39cc3c031242e6477acc1a4b9974cbba1dcad0d02cd9d156
3
  size 1074131104
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a3cf1c4501fb617a1ea6a6e7871b3840eb36338d12ad842dbbcf53c292df90d
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3e689b8aa4479620caaf93ab56a44aa0acb7f02b7879d1f2aa150ddbd93a61c
3
+ size 5240