Training in progress, epoch 1
Browse files
adapter_config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"alpha_pattern": {},
|
| 3 |
"auto_mapping": null,
|
| 4 |
-
"base_model_name_or_path": "
|
| 5 |
"bias": "none",
|
| 6 |
"fan_in_fan_out": false,
|
| 7 |
"inference_mode": true,
|
|
|
|
| 1 |
{
|
| 2 |
"alpha_pattern": {},
|
| 3 |
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 5 |
"bias": "none",
|
| 6 |
"fan_in_fan_out": false,
|
| 7 |
"inference_mode": true,
|
adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e1f52562d3242f155d0c0a6cae3f4a4b2571a7691e35d21ece164effa8353a1
|
| 3 |
+
size 27944408
|
runs/Jul22_11-42-24_39d74df35b21/events.out.tfevents.1721648617.39d74df35b21.617.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f4df343961199af7d466c9f2857af12d51d6a9d85f5e5bcf3fa836b969dfac47
|
| 3 |
+
size 6148
|
tokenizer.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer.model
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
| 3 |
+
size 493443
|
tokenizer_config.json
CHANGED
|
@@ -28,15 +28,16 @@
|
|
| 28 |
"special": true
|
| 29 |
}
|
| 30 |
},
|
|
|
|
| 31 |
"bos_token": "<s>",
|
| 32 |
-
"chat_template": "{
|
| 33 |
"clean_up_tokenization_spaces": false,
|
| 34 |
"eos_token": "</s>",
|
| 35 |
"legacy": false,
|
| 36 |
"model_max_length": 1000000000000000019884624838656,
|
| 37 |
"pad_token": "</s>",
|
| 38 |
-
"padding_side": "right",
|
| 39 |
"sp_model_kwargs": {},
|
|
|
|
| 40 |
"tokenizer_class": "LlamaTokenizer",
|
| 41 |
"unk_token": "<unk>",
|
| 42 |
"use_default_system_prompt": false
|
|
|
|
| 28 |
"special": true
|
| 29 |
}
|
| 30 |
},
|
| 31 |
+
"additional_special_tokens": [],
|
| 32 |
"bos_token": "<s>",
|
| 33 |
+
"chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\\n\\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + eos_token}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
|
| 34 |
"clean_up_tokenization_spaces": false,
|
| 35 |
"eos_token": "</s>",
|
| 36 |
"legacy": false,
|
| 37 |
"model_max_length": 1000000000000000019884624838656,
|
| 38 |
"pad_token": "</s>",
|
|
|
|
| 39 |
"sp_model_kwargs": {},
|
| 40 |
+
"spaces_between_special_tokens": false,
|
| 41 |
"tokenizer_class": "LlamaTokenizer",
|
| 42 |
"unk_token": "<unk>",
|
| 43 |
"use_default_system_prompt": false
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5432
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:498249626d0e8acf4e4d3e58fa5bcfe30a5538ae512404e97327375db5a88f40
|
| 3 |
size 5432
|