somya-kr commited on
Commit
46d525b
·
verified ·
1 Parent(s): 544fb17

End of training

Browse files
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: google/flan-t5-small
4
  tags:
@@ -36,9 +37,9 @@ The following hyperparameters were used during training:
36
  - train_batch_size: 8
37
  - eval_batch_size: 8
38
  - seed: 42
39
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
- - num_epochs: 2
42
 
43
  ### Training results
44
 
@@ -46,7 +47,7 @@ The following hyperparameters were used during training:
46
 
47
  ### Framework versions
48
 
49
- - Transformers 4.36.2
50
- - Pytorch 2.1.2+cu121
51
- - Datasets 2.16.1
52
- - Tokenizers 0.15.2
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: google/flan-t5-small
5
  tags:
 
37
  - train_batch_size: 8
38
  - eval_batch_size: 8
39
  - seed: 42
40
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
41
  - lr_scheduler_type: linear
42
+ - num_epochs: 10
43
 
44
  ### Training results
45
 
 
47
 
48
  ### Framework versions
49
 
50
+ - Transformers 4.46.1
51
+ - Pytorch 2.5.0+cu124
52
+ - Datasets 3.1.0
53
+ - Tokenizers 0.20.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2e927ea1b19cc324a97b48d48228febb5bbb5c3a8a692c88315b066ddb18447
3
  size 243128252
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9872397413d34b38322526da7648d65c03d9be2759b91a740e4c59757a8828d5
3
  size 243128252
runs/Nov04_13-56-10_raptor/events.out.tfevents.1730708777.raptor.1485474.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:286a69c9100a992e2d7574d4d4f2919f236f1a6912ffb34f88ee828b76b1a412
3
- size 42674
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ba4eafe8cbffa54ea3e9ddb4b2101f9ad789e3dfb8179151199faf257e03bb3
3
+ size 47119
tokenizer.json CHANGED
@@ -964,8 +964,8 @@
964
  "pre_tokenizer": {
965
  "type": "Metaspace",
966
  "replacement": "▁",
967
- "add_prefix_space": true,
968
- "prepend_scheme": "always"
969
  },
970
  "post_processor": {
971
  "type": "TemplateProcessing",
@@ -1024,8 +1024,8 @@
1024
  "decoder": {
1025
  "type": "Metaspace",
1026
  "replacement": "▁",
1027
- "add_prefix_space": true,
1028
- "prepend_scheme": "always"
1029
  },
1030
  "model": {
1031
  "type": "Unigram",
 
964
  "pre_tokenizer": {
965
  "type": "Metaspace",
966
  "replacement": "▁",
967
+ "prepend_scheme": "always",
968
+ "split": true
969
  },
970
  "post_processor": {
971
  "type": "TemplateProcessing",
 
1024
  "decoder": {
1025
  "type": "Metaspace",
1026
  "replacement": "▁",
1027
+ "prepend_scheme": "always",
1028
+ "split": true
1029
  },
1030
  "model": {
1031
  "type": "Unigram",
tokenizer_config.json CHANGED
@@ -927,7 +927,7 @@
927
  "<extra_id_98>",
928
  "<extra_id_99>"
929
  ],
930
- "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
  "model_max_length": 512,
 
927
  "<extra_id_98>",
928
  "<extra_id_99>"
929
  ],
930
+ "clean_up_tokenization_spaces": false,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
  "model_max_length": 512,