Training in progress, epoch 1
Browse files- adapter_config.json +3 -3
- adapter_model.safetensors +2 -2
- tokenizer.json +2 -35
- tokenizer_config.json +2 -2
- training_args.bin +1 -1
adapter_config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"alpha_pattern": {},
|
| 3 |
"auto_mapping": null,
|
| 4 |
-
"base_model_name_or_path": "
|
| 5 |
"bias": "none",
|
| 6 |
"fan_in_fan_out": false,
|
| 7 |
"inference_mode": true,
|
|
@@ -22,8 +22,8 @@
|
|
| 22 |
"rank_pattern": {},
|
| 23 |
"revision": null,
|
| 24 |
"target_modules": [
|
| 25 |
-
"
|
| 26 |
-
"
|
| 27 |
],
|
| 28 |
"task_type": "SEQ_CLASSIFICATION",
|
| 29 |
"use_dora": false,
|
|
|
|
| 1 |
{
|
| 2 |
"alpha_pattern": {},
|
| 3 |
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "microsoft/Phi-3-mini-128k-instruct",
|
| 5 |
"bias": "none",
|
| 6 |
"fan_in_fan_out": false,
|
| 7 |
"inference_mode": true,
|
|
|
|
| 22 |
"rank_pattern": {},
|
| 23 |
"revision": null,
|
| 24 |
"target_modules": [
|
| 25 |
+
"qkv_proj",
|
| 26 |
+
"o_proj"
|
| 27 |
],
|
| 28 |
"task_type": "SEQ_CLASSIFICATION",
|
| 29 |
"use_dora": false,
|
adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ccff1fa6c478364af95295064013f402a9ee242a16c23b35e2cf31e47e9d4991
|
| 3 |
+
size 18965320
|
tokenizer.json
CHANGED
|
@@ -1,11 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
-
"truncation":
|
| 4 |
-
"direction": "Right",
|
| 5 |
-
"max_length": 2048,
|
| 6 |
-
"strategy": "LongestFirst",
|
| 7 |
-
"stride": 0
|
| 8 |
-
},
|
| 9 |
"padding": null,
|
| 10 |
"added_tokens": [
|
| 11 |
{
|
|
@@ -155,12 +150,6 @@
|
|
| 155 |
"post_processor": {
|
| 156 |
"type": "TemplateProcessing",
|
| 157 |
"single": [
|
| 158 |
-
{
|
| 159 |
-
"SpecialToken": {
|
| 160 |
-
"id": "<s>",
|
| 161 |
-
"type_id": 0
|
| 162 |
-
}
|
| 163 |
-
},
|
| 164 |
{
|
| 165 |
"Sequence": {
|
| 166 |
"id": "A",
|
|
@@ -169,24 +158,12 @@
|
|
| 169 |
}
|
| 170 |
],
|
| 171 |
"pair": [
|
| 172 |
-
{
|
| 173 |
-
"SpecialToken": {
|
| 174 |
-
"id": "<s>",
|
| 175 |
-
"type_id": 0
|
| 176 |
-
}
|
| 177 |
-
},
|
| 178 |
{
|
| 179 |
"Sequence": {
|
| 180 |
"id": "A",
|
| 181 |
"type_id": 0
|
| 182 |
}
|
| 183 |
},
|
| 184 |
-
{
|
| 185 |
-
"SpecialToken": {
|
| 186 |
-
"id": "<s>",
|
| 187 |
-
"type_id": 1
|
| 188 |
-
}
|
| 189 |
-
},
|
| 190 |
{
|
| 191 |
"Sequence": {
|
| 192 |
"id": "B",
|
|
@@ -194,17 +171,7 @@
|
|
| 194 |
}
|
| 195 |
}
|
| 196 |
],
|
| 197 |
-
"special_tokens": {
|
| 198 |
-
"<s>": {
|
| 199 |
-
"id": "<s>",
|
| 200 |
-
"ids": [
|
| 201 |
-
1
|
| 202 |
-
],
|
| 203 |
-
"tokens": [
|
| 204 |
-
"<s>"
|
| 205 |
-
]
|
| 206 |
-
}
|
| 207 |
-
}
|
| 208 |
},
|
| 209 |
"decoder": {
|
| 210 |
"type": "Sequence",
|
|
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
+
"truncation": null,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
"padding": null,
|
| 5 |
"added_tokens": [
|
| 6 |
{
|
|
|
|
| 150 |
"post_processor": {
|
| 151 |
"type": "TemplateProcessing",
|
| 152 |
"single": [
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 153 |
{
|
| 154 |
"Sequence": {
|
| 155 |
"id": "A",
|
|
|
|
| 158 |
}
|
| 159 |
],
|
| 160 |
"pair": [
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 161 |
{
|
| 162 |
"Sequence": {
|
| 163 |
"id": "A",
|
| 164 |
"type_id": 0
|
| 165 |
}
|
| 166 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 167 |
{
|
| 168 |
"Sequence": {
|
| 169 |
"id": "B",
|
|
|
|
| 171 |
}
|
| 172 |
}
|
| 173 |
],
|
| 174 |
+
"special_tokens": {}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 175 |
},
|
| 176 |
"decoder": {
|
| 177 |
"type": "Sequence",
|
tokenizer_config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"add_bos_token":
|
| 3 |
"add_eos_token": false,
|
| 4 |
"add_prefix_space": null,
|
| 5 |
"added_tokens_decoder": {
|
|
@@ -117,7 +117,7 @@
|
|
| 117 |
}
|
| 118 |
},
|
| 119 |
"bos_token": "<s>",
|
| 120 |
-
"chat_template": "{
|
| 121 |
"clean_up_tokenization_spaces": false,
|
| 122 |
"eos_token": "<|endoftext|>",
|
| 123 |
"legacy": false,
|
|
|
|
| 1 |
{
|
| 2 |
+
"add_bos_token": false,
|
| 3 |
"add_eos_token": false,
|
| 4 |
"add_prefix_space": null,
|
| 5 |
"added_tokens_decoder": {
|
|
|
|
| 117 |
}
|
| 118 |
},
|
| 119 |
"bos_token": "<s>",
|
| 120 |
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}",
|
| 121 |
"clean_up_tokenization_spaces": false,
|
| 122 |
"eos_token": "<|endoftext|>",
|
| 123 |
"legacy": false,
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5176
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a2de8d932e0f2a109536714b25257659595a5dcd467de13d6c0c868c47b1d08
|
| 3 |
size 5176
|