minpeter commited on
Commit
53e79b0
·
verified ·
1 Parent(s): 6cf42d2

End of training

Browse files
Files changed (1) hide show
  1. README.md +23 -23
README.md CHANGED
@@ -14,7 +14,7 @@ datasets:
14
  - coastral/korean-writing-style-instruct
15
  - devngho/korean-instruction-mix
16
  model-index:
17
- - name: ko-tiny-exp
18
  results: []
19
  ---
20
 
@@ -28,6 +28,11 @@ axolotl version: `0.10.0.dev0`
28
  ```yaml
29
  base_model: minpeter/pretrained-tiny-ko
30
 
 
 
 
 
 
31
  chat_template: chatml
32
  datasets:
33
  - path: lemon-mint/Korean-FineTome-100k
@@ -102,11 +107,6 @@ datasets:
102
  dataset_prepared_path: last_run_prepared
103
  val_set_size: 0.05
104
 
105
- hub_model_id: minpeter/ko-tiny-exp
106
- output_dir: ./ouputs/ko-tiny-exp
107
- wandb_project: "axolotl"
108
- wandb_entity: "kasfiekfs-e"
109
-
110
  save_steps: 200
111
  warmup_steps: 20
112
  eval_steps: 200
@@ -153,11 +153,11 @@ weight_decay: 0.0
153
 
154
  </details><br>
155
 
156
- # ko-tiny-exp
157
 
158
  This model is a fine-tuned version of [minpeter/pretrained-tiny-ko](https://huggingface.co/minpeter/pretrained-tiny-ko) on the lemon-mint/Korean-FineTome-100k, the lemon-mint/smol-koreantalk, the heegyu/open-korean-instructions-v20231020, the FreedomIntelligence/evol-instruct-korean, the FreedomIntelligence/alpaca-gpt4-korean, the FreedomIntelligence/sharegpt-korean, the coastral/korean-writing-style-instruct and the devngho/korean-instruction-mix datasets.
159
  It achieves the following results on the evaluation set:
160
- - Loss: 1.4634
161
 
162
  ## Model description
163
 
@@ -194,21 +194,21 @@ The following hyperparameters were used during training:
194
 
195
  | Training Loss | Epoch | Step | Validation Loss |
196
  |:-------------:|:------:|:----:|:---------------:|
197
- | 2.696 | 0.0010 | 1 | 2.7432 |
198
- | 1.7677 | 0.2019 | 200 | 1.7528 |
199
- | 1.6696 | 0.4037 | 400 | 1.6833 |
200
- | 1.5866 | 0.6056 | 600 | 1.6401 |
201
- | 1.6249 | 0.8075 | 800 | 1.5957 |
202
- | 1.3578 | 1.0091 | 1000 | 1.5704 |
203
- | 1.4469 | 1.2110 | 1200 | 1.5514 |
204
- | 1.3969 | 1.4128 | 1400 | 1.5220 |
205
- | 1.3549 | 1.6147 | 1600 | 1.4939 |
206
- | 1.3107 | 1.8166 | 1800 | 1.4695 |
207
- | 1.2462 | 2.0182 | 2000 | 1.4751 |
208
- | 1.2001 | 2.2200 | 2200 | 1.4692 |
209
- | 1.0911 | 2.4219 | 2400 | 1.4661 |
210
- | 1.1547 | 2.6238 | 2600 | 1.4636 |
211
- | 1.1943 | 2.8256 | 2800 | 1.4634 |
212
 
213
 
214
  ### Framework versions
 
14
  - coastral/korean-writing-style-instruct
15
  - devngho/korean-instruction-mix
16
  model-index:
17
+ - name: tiny-ko-sft
18
  results: []
19
  ---
20
 
 
28
  ```yaml
29
  base_model: minpeter/pretrained-tiny-ko
30
 
31
+ hub_model_id: minpeter/tiny-ko-sft
32
+ output_dir: ./outputs/tiny-ko-sft
33
+ wandb_project: "axolotl"
34
+ wandb_entity: "kasfiekfs-e"
35
+
36
  chat_template: chatml
37
  datasets:
38
  - path: lemon-mint/Korean-FineTome-100k
 
107
  dataset_prepared_path: last_run_prepared
108
  val_set_size: 0.05
109
 
 
 
 
 
 
110
  save_steps: 200
111
  warmup_steps: 20
112
  eval_steps: 200
 
153
 
154
  </details><br>
155
 
156
+ # tiny-ko-sft
157
 
158
  This model is a fine-tuned version of [minpeter/pretrained-tiny-ko](https://huggingface.co/minpeter/pretrained-tiny-ko) on the lemon-mint/Korean-FineTome-100k, the lemon-mint/smol-koreantalk, the heegyu/open-korean-instructions-v20231020, the FreedomIntelligence/evol-instruct-korean, the FreedomIntelligence/alpaca-gpt4-korean, the FreedomIntelligence/sharegpt-korean, the coastral/korean-writing-style-instruct and the devngho/korean-instruction-mix datasets.
159
  It achieves the following results on the evaluation set:
160
+ - Loss: 1.4286
161
 
162
  ## Model description
163
 
 
194
 
195
  | Training Loss | Epoch | Step | Validation Loss |
196
  |:-------------:|:------:|:----:|:---------------:|
197
+ | 2.9956 | 0.0010 | 1 | 3.0182 |
198
+ | 1.7162 | 0.2019 | 200 | 1.7023 |
199
+ | 1.6186 | 0.4037 | 400 | 1.6351 |
200
+ | 1.5474 | 0.6056 | 600 | 1.5951 |
201
+ | 1.5822 | 0.8075 | 800 | 1.5540 |
202
+ | 1.3144 | 1.0091 | 1000 | 1.5333 |
203
+ | 1.403 | 1.2110 | 1200 | 1.5128 |
204
+ | 1.3558 | 1.4128 | 1400 | 1.4832 |
205
+ | 1.3165 | 1.6147 | 1600 | 1.4541 |
206
+ | 1.2704 | 1.8166 | 1800 | 1.4305 |
207
+ | 1.1913 | 2.0182 | 2000 | 1.4424 |
208
+ | 1.1488 | 2.2200 | 2200 | 1.4346 |
209
+ | 1.0417 | 2.4219 | 2400 | 1.4311 |
210
+ | 1.1104 | 2.6238 | 2600 | 1.4288 |
211
+ | 1.1446 | 2.8256 | 2800 | 1.4286 |
212
 
213
 
214
  ### Framework versions