TymofiiNasobko/Lapa-function-calling
Browse files- README.md +2 -2
- adapter_config.json +10 -7
- adapter_model.safetensors +2 -2
- chat_template.jinja +3 -3
- special_tokens_map.json +1 -1
- tokenizer_config.json +1 -1
- training_args.bin +2 -2
README.md
CHANGED
|
@@ -4,8 +4,8 @@ library_name: transformers
|
|
| 4 |
model_name: Lapa-function-calling
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
| 7 |
-
- trl
|
| 8 |
- sft
|
|
|
|
| 9 |
licence: license
|
| 10 |
---
|
| 11 |
|
|
@@ -36,7 +36,7 @@ This model was trained with SFT.
|
|
| 36 |
|
| 37 |
- TRL: 0.25.1
|
| 38 |
- Transformers: 4.57.3
|
| 39 |
-
- Pytorch: 2.
|
| 40 |
- Datasets: 4.4.1
|
| 41 |
- Tokenizers: 0.22.1
|
| 42 |
|
|
|
|
| 4 |
model_name: Lapa-function-calling
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
|
|
|
| 7 |
- sft
|
| 8 |
+
- trl
|
| 9 |
licence: license
|
| 10 |
---
|
| 11 |
|
|
|
|
| 36 |
|
| 37 |
- TRL: 0.25.1
|
| 38 |
- Transformers: 4.57.3
|
| 39 |
+
- Pytorch: 2.6.0+cu118
|
| 40 |
- Datasets: 4.4.1
|
| 41 |
- Tokenizers: 0.22.1
|
| 42 |
|
adapter_config.json
CHANGED
|
@@ -21,7 +21,10 @@
|
|
| 21 |
"lora_dropout": 0.05,
|
| 22 |
"megatron_config": null,
|
| 23 |
"megatron_core": "megatron.core",
|
| 24 |
-
"modules_to_save":
|
|
|
|
|
|
|
|
|
|
| 25 |
"peft_type": "LORA",
|
| 26 |
"peft_version": "0.18.0",
|
| 27 |
"qalora_group_size": 16,
|
|
@@ -29,16 +32,16 @@
|
|
| 29 |
"rank_pattern": {},
|
| 30 |
"revision": null,
|
| 31 |
"target_modules": [
|
| 32 |
-
"fc2",
|
| 33 |
"o_proj",
|
| 34 |
-
"
|
| 35 |
-
"
|
|
|
|
| 36 |
"q_proj",
|
|
|
|
| 37 |
"fc1",
|
| 38 |
"up_proj",
|
| 39 |
-
"
|
| 40 |
-
"
|
| 41 |
-
"v_proj"
|
| 42 |
],
|
| 43 |
"target_parameters": null,
|
| 44 |
"task_type": "CAUSAL_LM",
|
|
|
|
| 21 |
"lora_dropout": 0.05,
|
| 22 |
"megatron_config": null,
|
| 23 |
"megatron_core": "megatron.core",
|
| 24 |
+
"modules_to_save": [
|
| 25 |
+
"lm_head",
|
| 26 |
+
"embed_tokens"
|
| 27 |
+
],
|
| 28 |
"peft_type": "LORA",
|
| 29 |
"peft_version": "0.18.0",
|
| 30 |
"qalora_group_size": 16,
|
|
|
|
| 32 |
"rank_pattern": {},
|
| 33 |
"revision": null,
|
| 34 |
"target_modules": [
|
|
|
|
| 35 |
"o_proj",
|
| 36 |
+
"v_proj",
|
| 37 |
+
"down_proj",
|
| 38 |
+
"fc2",
|
| 39 |
"q_proj",
|
| 40 |
+
"out_proj",
|
| 41 |
"fc1",
|
| 42 |
"up_proj",
|
| 43 |
+
"gate_proj",
|
| 44 |
+
"k_proj"
|
|
|
|
| 45 |
],
|
| 46 |
"target_parameters": null,
|
| 47 |
"task_type": "CAUSAL_LM",
|
adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d5bd45b8610d7433d4628cc89c58265703031a5b658a642c73bf3bb7de3c0ffd
|
| 3 |
+
size 8774281960
|
chat_template.jinja
CHANGED
|
@@ -42,14 +42,14 @@
|
|
| 42 |
{%- else -%}
|
| 43 |
{{ raise_exception("Invalid content type") }}
|
| 44 |
{%- endif -%}
|
| 45 |
-
|
| 46 |
{% generation %}
|
| 47 |
{{ '<end_of_turn>
|
| 48 |
' }}{% endgeneration %}
|
| 49 |
-
|
| 50 |
{{ '<end_of_turn>
|
| 51 |
' }}
|
| 52 |
-
|
| 53 |
{%- endfor -%}
|
| 54 |
|
| 55 |
{%- if add_generation_prompt -%}
|
|
|
|
| 42 |
{%- else -%}
|
| 43 |
{{ raise_exception("Invalid content type") }}
|
| 44 |
{%- endif -%}
|
| 45 |
+
{%- if role == "model" -%}
|
| 46 |
{% generation %}
|
| 47 |
{{ '<end_of_turn>
|
| 48 |
' }}{% endgeneration %}
|
| 49 |
+
{%- else -%}
|
| 50 |
{{ '<end_of_turn>
|
| 51 |
' }}
|
| 52 |
+
{%- endif -%}
|
| 53 |
{%- endfor -%}
|
| 54 |
|
| 55 |
{%- if add_generation_prompt -%}
|
special_tokens_map.json
CHANGED
|
@@ -8,7 +8,7 @@
|
|
| 8 |
"single_word": false
|
| 9 |
},
|
| 10 |
"eoi_token": "<end_of_image>",
|
| 11 |
-
"eos_token": "<
|
| 12 |
"image_token": "<image_soft_token>",
|
| 13 |
"pad_token": {
|
| 14 |
"content": "<pad>",
|
|
|
|
| 8 |
"single_word": false
|
| 9 |
},
|
| 10 |
"eoi_token": "<end_of_image>",
|
| 11 |
+
"eos_token": "<end_of_turn>",
|
| 12 |
"image_token": "<image_soft_token>",
|
| 13 |
"pad_token": {
|
| 14 |
"content": "<pad>",
|
tokenizer_config.json
CHANGED
|
@@ -2207,7 +2207,7 @@
|
|
| 2207 |
"bos_token": "<bos>",
|
| 2208 |
"clean_up_tokenization_spaces": false,
|
| 2209 |
"eoi_token": "<end_of_image>",
|
| 2210 |
-
"eos_token": "<
|
| 2211 |
"extra_special_tokens": {
|
| 2212 |
"boi_token": "<start_of_image>",
|
| 2213 |
"eoi_token": "<end_of_image>",
|
|
|
|
| 2207 |
"bos_token": "<bos>",
|
| 2208 |
"clean_up_tokenization_spaces": false,
|
| 2209 |
"eoi_token": "<end_of_image>",
|
| 2210 |
+
"eos_token": "<end_of_turn>",
|
| 2211 |
"extra_special_tokens": {
|
| 2212 |
"boi_token": "<start_of_image>",
|
| 2213 |
"eoi_token": "<end_of_image>",
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:30e1c54445c5beccd239bb6cc38804b3a0439fb2dae5c02ee756d2bb97e018e9
|
| 3 |
+
size 5816
|