TymofiiNasobko commited on
Commit
defc7d7
·
verified ·
1 Parent(s): 5c14664

TymofiiNasobko/Lapa-function-calling

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: Lapa-function-calling
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - sft
 
9
  licence: license
10
  ---
11
 
@@ -36,7 +36,7 @@ This model was trained with SFT.
36
 
37
  - TRL: 0.25.1
38
  - Transformers: 4.57.3
39
- - Pytorch: 2.8.0+cu129
40
  - Datasets: 4.4.1
41
  - Tokenizers: 0.22.1
42
 
 
4
  model_name: Lapa-function-calling
5
  tags:
6
  - generated_from_trainer
 
7
  - sft
8
+ - trl
9
  licence: license
10
  ---
11
 
 
36
 
37
  - TRL: 0.25.1
38
  - Transformers: 4.57.3
39
+ - Pytorch: 2.6.0+cu118
40
  - Datasets: 4.4.1
41
  - Tokenizers: 0.22.1
42
 
adapter_config.json CHANGED
@@ -21,7 +21,10 @@
21
  "lora_dropout": 0.05,
22
  "megatron_config": null,
23
  "megatron_core": "megatron.core",
24
- "modules_to_save": null,
 
 
 
25
  "peft_type": "LORA",
26
  "peft_version": "0.18.0",
27
  "qalora_group_size": 16,
@@ -29,16 +32,16 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
- "fc2",
33
  "o_proj",
34
- "k_proj",
35
- "gate_proj",
 
36
  "q_proj",
 
37
  "fc1",
38
  "up_proj",
39
- "down_proj",
40
- "out_proj",
41
- "v_proj"
42
  ],
43
  "target_parameters": null,
44
  "task_type": "CAUSAL_LM",
 
21
  "lora_dropout": 0.05,
22
  "megatron_config": null,
23
  "megatron_core": "megatron.core",
24
+ "modules_to_save": [
25
+ "lm_head",
26
+ "embed_tokens"
27
+ ],
28
  "peft_type": "LORA",
29
  "peft_version": "0.18.0",
30
  "qalora_group_size": 16,
 
32
  "rank_pattern": {},
33
  "revision": null,
34
  "target_modules": [
 
35
  "o_proj",
36
+ "v_proj",
37
+ "down_proj",
38
+ "fc2",
39
  "q_proj",
40
+ "out_proj",
41
  "fc1",
42
  "up_proj",
43
+ "gate_proj",
44
+ "k_proj"
 
45
  ],
46
  "target_parameters": null,
47
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b51c582896766642a149c855746afbcb3818bd27cda03655a6cd4de69ae656e8
3
- size 4746766832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5bd45b8610d7433d4628cc89c58265703031a5b658a642c73bf3bb7de3c0ffd
3
+ size 8774281960
chat_template.jinja CHANGED
@@ -42,14 +42,14 @@
42
  {%- else -%}
43
  {{ raise_exception("Invalid content type") }}
44
  {%- endif -%}
45
- {%- if role == "model" -%}
46
  {% generation %}
47
  {{ '<end_of_turn>
48
  ' }}{% endgeneration %}
49
- {%- else -%}
50
  {{ '<end_of_turn>
51
  ' }}
52
- {%- endif -%}
53
  {%- endfor -%}
54
 
55
  {%- if add_generation_prompt -%}
 
42
  {%- else -%}
43
  {{ raise_exception("Invalid content type") }}
44
  {%- endif -%}
45
+ {%- if role == "model" -%}
46
  {% generation %}
47
  {{ '<end_of_turn>
48
  ' }}{% endgeneration %}
49
+ {%- else -%}
50
  {{ '<end_of_turn>
51
  ' }}
52
+ {%- endif -%}
53
  {%- endfor -%}
54
 
55
  {%- if add_generation_prompt -%}
special_tokens_map.json CHANGED
@@ -8,7 +8,7 @@
8
  "single_word": false
9
  },
10
  "eoi_token": "<end_of_image>",
11
- "eos_token": "<eos>",
12
  "image_token": "<image_soft_token>",
13
  "pad_token": {
14
  "content": "<pad>",
 
8
  "single_word": false
9
  },
10
  "eoi_token": "<end_of_image>",
11
+ "eos_token": "<end_of_turn>",
12
  "image_token": "<image_soft_token>",
13
  "pad_token": {
14
  "content": "<pad>",
tokenizer_config.json CHANGED
@@ -2207,7 +2207,7 @@
2207
  "bos_token": "<bos>",
2208
  "clean_up_tokenization_spaces": false,
2209
  "eoi_token": "<end_of_image>",
2210
- "eos_token": "<eos>",
2211
  "extra_special_tokens": {
2212
  "boi_token": "<start_of_image>",
2213
  "eoi_token": "<end_of_image>",
 
2207
  "bos_token": "<bos>",
2208
  "clean_up_tokenization_spaces": false,
2209
  "eoi_token": "<end_of_image>",
2210
+ "eos_token": "<end_of_turn>",
2211
  "extra_special_tokens": {
2212
  "boi_token": "<start_of_image>",
2213
  "eoi_token": "<end_of_image>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:264c88be027ded39b70324149dbe1cf4b50e33e90c1a743861991a753109baff
3
- size 6225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30e1c54445c5beccd239bb6cc38804b3a0439fb2dae5c02ee756d2bb97e018e9
3
+ size 5816