SangMoone commited on
Commit
a147579
·
verified ·
1 Parent(s): 07f575b

SangMoone/beomi-ko-gemma_train_test_33000

Browse files
Files changed (31) hide show
  1. .gitattributes +1 -0
  2. README.md +189 -0
  3. adapter_config.json +29 -0
  4. adapter_model.safetensors +3 -0
  5. runs/Oct11_17-15-18_DESKTOP-SGQV3FH/events.out.tfevents.1728634637.DESKTOP-SGQV3FH.2637.0 +3 -0
  6. runs/Oct11_17-15-18_DESKTOP-SGQV3FH/events.out.tfevents.1728635032.DESKTOP-SGQV3FH.2637.1 +3 -0
  7. runs/Oct11_17-15-18_DESKTOP-SGQV3FH/events.out.tfevents.1728635420.DESKTOP-SGQV3FH.2637.2 +3 -0
  8. runs/Oct11_17-40-39_DESKTOP-SGQV3FH/events.out.tfevents.1728636080.DESKTOP-SGQV3FH.825.0 +3 -0
  9. runs/Oct11_17-44-53_DESKTOP-SGQV3FH/events.out.tfevents.1728636317.DESKTOP-SGQV3FH.825.1 +3 -0
  10. runs/Oct11_17-54-09_DESKTOP-SGQV3FH/events.out.tfevents.1728636877.DESKTOP-SGQV3FH.1113.0 +3 -0
  11. runs/Oct11_18-04-38_DESKTOP-SGQV3FH/events.out.tfevents.1728637536.DESKTOP-SGQV3FH.1269.0 +3 -0
  12. runs/Oct11_18-09-22_DESKTOP-SGQV3FH/events.out.tfevents.1728637799.DESKTOP-SGQV3FH.1269.1 +3 -0
  13. runs/Oct11_18-11-33_DESKTOP-SGQV3FH/events.out.tfevents.1728637925.DESKTOP-SGQV3FH.1467.0 +3 -0
  14. runs/Oct11_18-14-04_DESKTOP-SGQV3FH/events.out.tfevents.1728638072.DESKTOP-SGQV3FH.1467.1 +3 -0
  15. runs/Oct11_18-14-04_DESKTOP-SGQV3FH/events.out.tfevents.1728638132.DESKTOP-SGQV3FH.1467.2 +3 -0
  16. runs/Oct11_18-22-08_DESKTOP-SGQV3FH/events.out.tfevents.1728638554.DESKTOP-SGQV3FH.1837.0 +3 -0
  17. runs/Oct11_18-23-00_DESKTOP-SGQV3FH/events.out.tfevents.1728638605.DESKTOP-SGQV3FH.1837.1 +3 -0
  18. runs/Oct11_18-28-25_DESKTOP-SGQV3FH/events.out.tfevents.1728638929.DESKTOP-SGQV3FH.2704.0 +3 -0
  19. runs/Oct11_18-36-53_DESKTOP-SGQV3FH/events.out.tfevents.1728639455.DESKTOP-SGQV3FH.2704.1 +3 -0
  20. runs/Oct11_18-40-37_DESKTOP-SGQV3FH/events.out.tfevents.1728639662.DESKTOP-SGQV3FH.2912.0 +3 -0
  21. runs/Oct11_19-06-01_DESKTOP-SGQV3FH/events.out.tfevents.1728641190.DESKTOP-SGQV3FH.3219.0 +3 -0
  22. runs/Oct11_19-18-58_DESKTOP-SGQV3FH/events.out.tfevents.1728641982.DESKTOP-SGQV3FH.3392.0 +3 -0
  23. runs/Oct11_19-36-56_DESKTOP-SGQV3FH/events.out.tfevents.1728643027.DESKTOP-SGQV3FH.3642.0 +3 -0
  24. runs/Oct11_19-44-46_DESKTOP-SGQV3FH/events.out.tfevents.1728643503.DESKTOP-SGQV3FH.3838.0 +3 -0
  25. runs/Oct11_20-38-07_DESKTOP-SGQV3FH/events.out.tfevents.1728646701.DESKTOP-SGQV3FH.4336.0 +3 -0
  26. runs/Oct11_20-40-05_DESKTOP-SGQV3FH/events.out.tfevents.1728646815.DESKTOP-SGQV3FH.4505.0 +3 -0
  27. special_tokens_map.json +24 -0
  28. tokenizer.json +3 -0
  29. tokenizer.model +3 -0
  30. tokenizer_config.json +49 -0
  31. training_args.bin +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,189 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: beomi/gemma-ko-2b
3
+ library_name: peft
4
+ license: other
5
+ tags:
6
+ - trl
7
+ - sft
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: results_1011
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # results_1011
18
+
19
+ This model is a fine-tuned version of [beomi/gemma-ko-2b](https://huggingface.co/beomi/gemma-ko-2b) on the None dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 1.5227
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 0.0002
41
+ - train_batch_size: 3
42
+ - eval_batch_size: 2
43
+ - seed: 42
44
+ - gradient_accumulation_steps: 8
45
+ - total_train_batch_size: 24
46
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
+ - lr_scheduler_type: cosine
48
+ - lr_scheduler_warmup_ratio: 0.03
49
+ - num_epochs: 10
50
+ - mixed_precision_training: Native AMP
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss |
55
+ |:-------------:|:------:|:-----:|:---------------:|
56
+ | 2.0356 | 0.08 | 100 | 1.8434 |
57
+ | 1.7068 | 0.16 | 200 | 1.6624 |
58
+ | 1.6553 | 0.24 | 300 | 1.6477 |
59
+ | 1.6456 | 0.32 | 400 | 1.6441 |
60
+ | 1.6403 | 0.4 | 500 | 1.6378 |
61
+ | 1.6347 | 0.48 | 600 | 1.6315 |
62
+ | 1.6218 | 0.56 | 700 | 1.6256 |
63
+ | 1.6259 | 0.64 | 800 | 1.6192 |
64
+ | 1.6156 | 0.72 | 900 | 1.6116 |
65
+ | 1.6202 | 0.8 | 1000 | 1.6075 |
66
+ | 1.6031 | 0.88 | 1100 | 1.6058 |
67
+ | 1.6018 | 0.96 | 1200 | 1.6031 |
68
+ | 1.5965 | 1.04 | 1300 | 1.6022 |
69
+ | 1.5988 | 1.12 | 1400 | 1.6002 |
70
+ | 1.6043 | 1.2 | 1500 | 1.5978 |
71
+ | 1.5933 | 1.28 | 1600 | 1.5962 |
72
+ | 1.5909 | 1.3600 | 1700 | 1.5953 |
73
+ | 1.6014 | 1.44 | 1800 | 1.5932 |
74
+ | 1.584 | 1.52 | 1900 | 1.5912 |
75
+ | 1.5865 | 1.6 | 2000 | 1.5897 |
76
+ | 1.5871 | 1.6800 | 2100 | 1.5880 |
77
+ | 1.5838 | 1.76 | 2200 | 1.5865 |
78
+ | 1.5878 | 1.8400 | 2300 | 1.5850 |
79
+ | 1.58 | 1.92 | 2400 | 1.5835 |
80
+ | 1.5819 | 2.0 | 2500 | 1.5812 |
81
+ | 1.5652 | 2.08 | 2600 | 1.5806 |
82
+ | 1.573 | 2.16 | 2700 | 1.5796 |
83
+ | 1.5677 | 2.24 | 2800 | 1.5779 |
84
+ | 1.572 | 2.32 | 2900 | 1.5764 |
85
+ | 1.5688 | 2.4 | 3000 | 1.5748 |
86
+ | 1.5663 | 2.48 | 3100 | 1.5730 |
87
+ | 1.5669 | 2.56 | 3200 | 1.5719 |
88
+ | 1.5613 | 2.64 | 3300 | 1.5704 |
89
+ | 1.564 | 2.7200 | 3400 | 1.5690 |
90
+ | 1.5619 | 2.8 | 3500 | 1.5681 |
91
+ | 1.5622 | 2.88 | 3600 | 1.5667 |
92
+ | 1.5628 | 2.96 | 3700 | 1.5651 |
93
+ | 1.5514 | 3.04 | 3800 | 1.5645 |
94
+ | 1.5597 | 3.12 | 3900 | 1.5628 |
95
+ | 1.5499 | 3.2 | 4000 | 1.5622 |
96
+ | 1.5436 | 3.2800 | 4100 | 1.5610 |
97
+ | 1.5521 | 3.36 | 4200 | 1.5598 |
98
+ | 1.5389 | 3.44 | 4300 | 1.5585 |
99
+ | 1.5518 | 3.52 | 4400 | 1.5577 |
100
+ | 1.545 | 3.6 | 4500 | 1.5559 |
101
+ | 1.5383 | 3.68 | 4600 | 1.5552 |
102
+ | 1.5338 | 3.76 | 4700 | 1.5538 |
103
+ | 1.5452 | 3.84 | 4800 | 1.5522 |
104
+ | 1.5269 | 3.92 | 4900 | 1.5516 |
105
+ | 1.5342 | 4.0 | 5000 | 1.5507 |
106
+ | 1.5243 | 4.08 | 5100 | 1.5503 |
107
+ | 1.5209 | 4.16 | 5200 | 1.5498 |
108
+ | 1.5337 | 4.24 | 5300 | 1.5487 |
109
+ | 1.5261 | 4.32 | 5400 | 1.5477 |
110
+ | 1.5255 | 4.4 | 5500 | 1.5463 |
111
+ | 1.5342 | 4.48 | 5600 | 1.5459 |
112
+ | 1.5211 | 4.5600 | 5700 | 1.5447 |
113
+ | 1.5293 | 4.64 | 5800 | 1.5441 |
114
+ | 1.5203 | 4.72 | 5900 | 1.5425 |
115
+ | 1.5171 | 4.8 | 6000 | 1.5421 |
116
+ | 1.5239 | 4.88 | 6100 | 1.5412 |
117
+ | 1.5184 | 4.96 | 6200 | 1.5404 |
118
+ | 1.508 | 5.04 | 6300 | 1.5405 |
119
+ | 1.5113 | 5.12 | 6400 | 1.5396 |
120
+ | 1.5035 | 5.2 | 6500 | 1.5385 |
121
+ | 1.5082 | 5.28 | 6600 | 1.5380 |
122
+ | 1.5144 | 5.36 | 6700 | 1.5376 |
123
+ | 1.5052 | 5.44 | 6800 | 1.5367 |
124
+ | 1.5096 | 5.52 | 6900 | 1.5358 |
125
+ | 1.5139 | 5.6 | 7000 | 1.5348 |
126
+ | 1.5026 | 5.68 | 7100 | 1.5344 |
127
+ | 1.5061 | 5.76 | 7200 | 1.5339 |
128
+ | 1.5073 | 5.84 | 7300 | 1.5332 |
129
+ | 1.5082 | 5.92 | 7400 | 1.5323 |
130
+ | 1.5118 | 6.0 | 7500 | 1.5320 |
131
+ | 1.4939 | 6.08 | 7600 | 1.5323 |
132
+ | 1.4986 | 6.16 | 7700 | 1.5322 |
133
+ | 1.492 | 6.24 | 7800 | 1.5324 |
134
+ | 1.4889 | 6.32 | 7900 | 1.5309 |
135
+ | 1.4986 | 6.4 | 8000 | 1.5301 |
136
+ | 1.5003 | 6.48 | 8100 | 1.5297 |
137
+ | 1.5059 | 6.5600 | 8200 | 1.5295 |
138
+ | 1.4961 | 6.64 | 8300 | 1.5291 |
139
+ | 1.4938 | 6.72 | 8400 | 1.5279 |
140
+ | 1.5039 | 6.8 | 8500 | 1.5276 |
141
+ | 1.4892 | 6.88 | 8600 | 1.5272 |
142
+ | 1.5 | 6.96 | 8700 | 1.5268 |
143
+ | 1.4944 | 7.04 | 8800 | 1.5270 |
144
+ | 1.4941 | 7.12 | 8900 | 1.5265 |
145
+ | 1.4849 | 7.2 | 9000 | 1.5270 |
146
+ | 1.4924 | 7.28 | 9100 | 1.5261 |
147
+ | 1.4903 | 7.36 | 9200 | 1.5256 |
148
+ | 1.4909 | 7.44 | 9300 | 1.5254 |
149
+ | 1.4884 | 7.52 | 9400 | 1.5253 |
150
+ | 1.4874 | 7.6 | 9500 | 1.5253 |
151
+ | 1.4973 | 7.68 | 9600 | 1.5251 |
152
+ | 1.4835 | 7.76 | 9700 | 1.5247 |
153
+ | 1.4844 | 7.84 | 9800 | 1.5245 |
154
+ | 1.4845 | 7.92 | 9900 | 1.5242 |
155
+ | 1.4857 | 8.0 | 10000 | 1.5239 |
156
+ | 1.483 | 8.08 | 10100 | 1.5241 |
157
+ | 1.4875 | 8.16 | 10200 | 1.5238 |
158
+ | 1.488 | 8.24 | 10300 | 1.5238 |
159
+ | 1.4816 | 8.32 | 10400 | 1.5236 |
160
+ | 1.4887 | 8.4 | 10500 | 1.5233 |
161
+ | 1.4785 | 8.48 | 10600 | 1.5236 |
162
+ | 1.4802 | 8.56 | 10700 | 1.5232 |
163
+ | 1.4846 | 8.64 | 10800 | 1.5231 |
164
+ | 1.4832 | 8.72 | 10900 | 1.5231 |
165
+ | 1.4821 | 8.8 | 11000 | 1.5229 |
166
+ | 1.4837 | 8.88 | 11100 | 1.5230 |
167
+ | 1.4865 | 8.96 | 11200 | 1.5229 |
168
+ | 1.4855 | 9.04 | 11300 | 1.5228 |
169
+ | 1.4841 | 9.12 | 11400 | 1.5229 |
170
+ | 1.4765 | 9.2 | 11500 | 1.5230 |
171
+ | 1.4795 | 9.28 | 11600 | 1.5228 |
172
+ | 1.4848 | 9.36 | 11700 | 1.5228 |
173
+ | 1.4827 | 9.44 | 11800 | 1.5229 |
174
+ | 1.4883 | 9.52 | 11900 | 1.5228 |
175
+ | 1.4796 | 9.6 | 12000 | 1.5228 |
176
+ | 1.4899 | 9.68 | 12100 | 1.5228 |
177
+ | 1.4852 | 9.76 | 12200 | 1.5227 |
178
+ | 1.48 | 9.84 | 12300 | 1.5227 |
179
+ | 1.4823 | 9.92 | 12400 | 1.5227 |
180
+ | 1.4796 | 10.0 | 12500 | 1.5227 |
181
+
182
+
183
+ ### Framework versions
184
+
185
+ - PEFT 0.12.0
186
+ - Transformers 4.45.0
187
+ - Pytorch 2.4.0+cu121
188
+ - Datasets 2.21.0
189
+ - Tokenizers 0.20.1
adapter_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "beomi/gemma-ko-2b",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0.1,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 32,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
+ ],
26
+ "task_type": "CAUSAL_LM",
27
+ "use_dora": false,
28
+ "use_rslora": false
29
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8eaa1c1ff0dc122411bf90663a4592be0e3a5bc8b8236d4e07144ef91015fe39
3
+ size 14755192
runs/Oct11_17-15-18_DESKTOP-SGQV3FH/events.out.tfevents.1728634637.DESKTOP-SGQV3FH.2637.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7e660210ddd6fea07d826bd6a6b7a0d073b4a87a0ea6222c0819bdce9c2822e
3
+ size 13996
runs/Oct11_17-15-18_DESKTOP-SGQV3FH/events.out.tfevents.1728635032.DESKTOP-SGQV3FH.2637.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf4549189fa8185d43d9c120c8d536668df43d497a482df7423cf17c96f3f7e4
3
+ size 4724
runs/Oct11_17-15-18_DESKTOP-SGQV3FH/events.out.tfevents.1728635420.DESKTOP-SGQV3FH.2637.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e2a78c612dc172caf1aa5b64567e6352188e0cba37c2faaefa7a48f6aa33100
3
+ size 4724
runs/Oct11_17-40-39_DESKTOP-SGQV3FH/events.out.tfevents.1728636080.DESKTOP-SGQV3FH.825.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8be41e4bf00ab927f2c5d0f3784b0837e6af30271d0d7f09fb2deac36cbd6c0d
3
+ size 5284
runs/Oct11_17-44-53_DESKTOP-SGQV3FH/events.out.tfevents.1728636317.DESKTOP-SGQV3FH.825.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:631dcdff7689c3103b5ee4c4574fd4d50d7197a35466ce297e3f824a41a0f763
3
+ size 5284
runs/Oct11_17-54-09_DESKTOP-SGQV3FH/events.out.tfevents.1728636877.DESKTOP-SGQV3FH.1113.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4016eee07dfb9b10c7d11a386ebeade3af5f36f5340506aa961184a2d8340803
3
+ size 5284
runs/Oct11_18-04-38_DESKTOP-SGQV3FH/events.out.tfevents.1728637536.DESKTOP-SGQV3FH.1269.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0beba0ec38de0469ded95fe2647b7331ed47d4e2023e3aa80a35484267d3ee85
3
+ size 5284
runs/Oct11_18-09-22_DESKTOP-SGQV3FH/events.out.tfevents.1728637799.DESKTOP-SGQV3FH.1269.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfa344474cab29f5c0db44a6dccde85e21a11d264a2e4cc39cc82fc56e65309d
3
+ size 5284
runs/Oct11_18-11-33_DESKTOP-SGQV3FH/events.out.tfevents.1728637925.DESKTOP-SGQV3FH.1467.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cbd85204903afec38e65a8b5b31af5a448a3d5993222037d2070f96cda76604
3
+ size 5284
runs/Oct11_18-14-04_DESKTOP-SGQV3FH/events.out.tfevents.1728638072.DESKTOP-SGQV3FH.1467.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74e9589bcd4f5ee96945f00e7aeced5d0e0e5e55cf5cc513978915273fca0098
3
+ size 5283
runs/Oct11_18-14-04_DESKTOP-SGQV3FH/events.out.tfevents.1728638132.DESKTOP-SGQV3FH.1467.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d27e0198cc03d49e8ecd2e2a809521a2c0274f1f4c0f01bc9314b7f35911cd01
3
+ size 5283
runs/Oct11_18-22-08_DESKTOP-SGQV3FH/events.out.tfevents.1728638554.DESKTOP-SGQV3FH.1837.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:038c3cfb28b33d60fa9700775db155094a530b4c8f715dfa94708d2856a762aa
3
+ size 5286
runs/Oct11_18-23-00_DESKTOP-SGQV3FH/events.out.tfevents.1728638605.DESKTOP-SGQV3FH.1837.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d157ee45f4283f4c489c2d4c04928ba6eefc04f63d1fff1503a0ee28da022c2
3
+ size 5285
runs/Oct11_18-28-25_DESKTOP-SGQV3FH/events.out.tfevents.1728638929.DESKTOP-SGQV3FH.2704.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f60f928c07932cf89332842ea49693b97a7c2e9409c868767e0f459b040889f
3
+ size 5492
runs/Oct11_18-36-53_DESKTOP-SGQV3FH/events.out.tfevents.1728639455.DESKTOP-SGQV3FH.2704.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69917949eeee2690420c4a1f4a6a76591e940df67fdbf18a67584600e0c28051
3
+ size 5285
runs/Oct11_18-40-37_DESKTOP-SGQV3FH/events.out.tfevents.1728639662.DESKTOP-SGQV3FH.2912.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5afb505723c345c0f3bfb104f4426e55b75b00ff433c41a92057408aad3c0420
3
+ size 5492
runs/Oct11_19-06-01_DESKTOP-SGQV3FH/events.out.tfevents.1728641190.DESKTOP-SGQV3FH.3219.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75526522eba4708d0c2721feb929a12016fd74e50f981378a4392a9c5922004b
3
+ size 5492
runs/Oct11_19-18-58_DESKTOP-SGQV3FH/events.out.tfevents.1728641982.DESKTOP-SGQV3FH.3392.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d85a89f1b5441ada82b9bcf6bcd7d47f1a98f6426be7a970300604c9bc5f4eb
3
+ size 5492
runs/Oct11_19-36-56_DESKTOP-SGQV3FH/events.out.tfevents.1728643027.DESKTOP-SGQV3FH.3642.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2acf00b32c6a50f2f0146db821af35ac9621217a65c4facd49b87054a1f5cf60
3
+ size 5492
runs/Oct11_19-44-46_DESKTOP-SGQV3FH/events.out.tfevents.1728643503.DESKTOP-SGQV3FH.3838.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ae10d7ae7dfacad954896f71209db501d21c436c42a6a2f7462b55794dff4e3
3
+ size 11753
runs/Oct11_20-38-07_DESKTOP-SGQV3FH/events.out.tfevents.1728646701.DESKTOP-SGQV3FH.4336.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1c0366d3d97309fc680f1a9aa52fb4c0e0d88cca106da7f59b7fe3628086009
3
+ size 10482
runs/Oct11_20-40-05_DESKTOP-SGQV3FH/events.out.tfevents.1728646815.DESKTOP-SGQV3FH.4505.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fccdc2fff250b9bfa40b888563a412884bbcd479176f22e45211f68b486e83b
3
+ size 65880
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<bos>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<eos>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<eos>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25061472d22ae7fe55a4f545eedc096c94c364d2a40de80538684d48333959ac
3
+ size 34315195
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61a7b147390c64585d6c3543dd6fc636906c9af3865a5548f27f31aee1d4c8e2
3
+ size 4241003
tokenizer_config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<eos>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "<bos>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "<unk>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ }
37
+ },
38
+ "bos_token": "<bos>",
39
+ "clean_up_tokenization_spaces": false,
40
+ "eos_token": "<eos>",
41
+ "legacy": null,
42
+ "model_max_length": 1000000000000000019884624838656,
43
+ "pad_token": "<eos>",
44
+ "sp_model_kwargs": {},
45
+ "spaces_between_special_tokens": false,
46
+ "tokenizer_class": "GemmaTokenizer",
47
+ "unk_token": "<unk>",
48
+ "use_default_system_prompt": false
49
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:336491ff06469a93bcf2f58ab5ca81d955dc3d2cfd220c49f3944f981ee825f7
3
+ size 5432