minpeter commited on
Commit
a2c65f3
·
verified ·
1 Parent(s): 60f3e9b

End of training

Browse files
Files changed (1) hide show
  1. README.md +57 -6
README.md CHANGED
@@ -7,7 +7,12 @@ tags:
7
  datasets:
8
  - lemon-mint/Korean-FineTome-100k
9
  - lemon-mint/smol-koreantalk
 
 
10
  - FreedomIntelligence/alpaca-gpt4-korean
 
 
 
11
  model-index:
12
  - name: ko-tiny-exp
13
  results: []
@@ -41,6 +46,27 @@ datasets:
41
  role: role
42
  content: content
43
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
  - path: FreedomIntelligence/alpaca-gpt4-korean
45
  type: chat_template
46
  split: train[:20%]
@@ -49,6 +75,30 @@ datasets:
49
  role: from
50
  content: value
51
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
  dataset_prepared_path: last_run_prepared
53
  val_set_size: 0.05
54
 
@@ -80,7 +130,7 @@ added_tokens_overrides:
80
  128002: "<|im_start|>"
81
 
82
  special_tokens:
83
- bos_token: <|im_start|>
84
  eos_token: <|im_end|>
85
  pad_token: <|im_end|>
86
 
@@ -91,7 +141,7 @@ resume_from_checkpoint:
91
  logging_steps: 1
92
  flash_attention: true
93
 
94
- num_epochs: 4
95
  weight_decay: 0.0
96
 
97
  ```
@@ -100,9 +150,9 @@ weight_decay: 0.0
100
 
101
  # ko-tiny-exp
102
 
103
- This model is a fine-tuned version of [minpeter/pretrained-tiny-ko](https://huggingface.co/minpeter/pretrained-tiny-ko) on the lemon-mint/Korean-FineTome-100k, the lemon-mint/smol-koreantalk and the FreedomIntelligence/alpaca-gpt4-korean datasets.
104
  It achieves the following results on the evaluation set:
105
- - Loss: 3.5174
106
 
107
  ## Model description
108
 
@@ -133,13 +183,14 @@ The following hyperparameters were used during training:
133
  - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
134
  - lr_scheduler_type: cosine
135
  - lr_scheduler_warmup_steps: 100
136
- - training_steps: 112
137
 
138
  ### Training results
139
 
140
  | Training Loss | Epoch | Step | Validation Loss |
141
  |:-------------:|:------:|:----:|:---------------:|
142
- | 3.5354 | 0.0351 | 1 | 3.5174 |
 
143
 
144
 
145
  ### Framework versions
 
7
  datasets:
8
  - lemon-mint/Korean-FineTome-100k
9
  - lemon-mint/smol-koreantalk
10
+ - heegyu/open-korean-instructions-v20231020
11
+ - FreedomIntelligence/evol-instruct-korean
12
  - FreedomIntelligence/alpaca-gpt4-korean
13
+ - FreedomIntelligence/sharegpt-korean
14
+ - coastral/korean-writing-style-instruct
15
+ - devngho/korean-instruction-mix
16
  model-index:
17
  - name: ko-tiny-exp
18
  results: []
 
46
  role: role
47
  content: content
48
 
49
+ - path: heegyu/open-korean-instructions-v20231020
50
+ type: chat_template
51
+ split: train[:20%]
52
+ field_messages: conversations
53
+ message_property_mappings:
54
+ role: from
55
+ content: value
56
+ roles:
57
+ user: ["human", "user"]
58
+ assistant: ["gpt", "assistant", "bot"]
59
+ system: ["system", "input"]
60
+
61
+ # NOTE: https://github.com/FreedomIntelligence/MultilingualSIFT
62
+ - path: FreedomIntelligence/evol-instruct-korean
63
+ type: chat_template
64
+ split: train[:20%]
65
+ field_messages: conversations
66
+ message_property_mappings:
67
+ role: from
68
+ content: value
69
+
70
  - path: FreedomIntelligence/alpaca-gpt4-korean
71
  type: chat_template
72
  split: train[:20%]
 
75
  role: from
76
  content: value
77
 
78
+ - path: FreedomIntelligence/sharegpt-korean
79
+ type: chat_template
80
+ split: train[:20%]
81
+ field_messages: conversations
82
+ message_property_mappings:
83
+ role: from
84
+ content: value
85
+
86
+ - path: coastral/korean-writing-style-instruct
87
+ type: chat_template
88
+ split: train[:20%]
89
+ field_messages: conversations
90
+ message_property_mappings:
91
+ role: from
92
+ content: value
93
+
94
+ - path: devngho/korean-instruction-mix
95
+ type: chat_template
96
+ split: train[:20%]
97
+ field_messages: messages
98
+ message_property_mappings:
99
+ role: from
100
+ content: value
101
+
102
  dataset_prepared_path: last_run_prepared
103
  val_set_size: 0.05
104
 
 
130
  128002: "<|im_start|>"
131
 
132
  special_tokens:
133
+ bos_token: <|begin_of_text|>
134
  eos_token: <|im_end|>
135
  pad_token: <|im_end|>
136
 
 
141
  logging_steps: 1
142
  flash_attention: true
143
 
144
+ num_epochs: 3
145
  weight_decay: 0.0
146
 
147
  ```
 
150
 
151
  # ko-tiny-exp
152
 
153
+ This model is a fine-tuned version of [minpeter/pretrained-tiny-ko](https://huggingface.co/minpeter/pretrained-tiny-ko) on the lemon-mint/Korean-FineTome-100k, the lemon-mint/smol-koreantalk, the heegyu/open-korean-instructions-v20231020, the FreedomIntelligence/evol-instruct-korean, the FreedomIntelligence/alpaca-gpt4-korean, the FreedomIntelligence/sharegpt-korean, the coastral/korean-writing-style-instruct and the devngho/korean-instruction-mix datasets.
154
  It achieves the following results on the evaluation set:
155
+ - Loss: 2.0944
156
 
157
  ## Model description
158
 
 
183
  - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
184
  - lr_scheduler_type: cosine
185
  - lr_scheduler_warmup_steps: 100
186
+ - training_steps: 264
187
 
188
  ### Training results
189
 
190
  | Training Loss | Epoch | Step | Validation Loss |
191
  |:-------------:|:------:|:----:|:---------------:|
192
+ | 3.362 | 0.0114 | 1 | 3.3719 |
193
+ | 2.1121 | 2.2727 | 200 | 2.0944 |
194
 
195
 
196
  ### Framework versions