Baileyyeah commited on
Commit
8c424a0
·
verified ·
1 Parent(s): 00ab8a4

Add files using upload-large-folder tool

Browse files
Files changed (49) hide show
  1. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/README.md +67 -0
  2. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/added_tokens.json +24 -0
  3. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/chat_template.jinja +54 -0
  4. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/config.json +66 -0
  5. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/merges.txt +0 -0
  6. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/model.safetensors.index.json +443 -0
  7. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/special_tokens_map.json +31 -0
  8. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/tokenizer_config.json +208 -0
  9. Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/trainer_state.json +2331 -0
  10. Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/added_tokens.json +24 -0
  11. Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/chat_template.jinja +54 -0
  12. Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/config.json +66 -0
  13. Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json +14 -0
  14. Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/model.safetensors.index.json +443 -0
  15. Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/special_tokens_map.json +31 -0
  16. Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/tokenizer_config.json +208 -0
  17. Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/added_tokens.json +24 -0
  18. Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/chat_template.jinja +54 -0
  19. Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/config.json +66 -0
  20. Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json +14 -0
  21. Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/model.safetensors.index.json +443 -0
  22. Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/tokenizer_config.json +208 -0
  23. Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/README.md +67 -0
  24. Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/added_tokens.json +24 -0
  25. Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/chat_template.jinja +54 -0
  26. Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/config.json +66 -0
  27. Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/generation_config.json +14 -0
  28. Blood/seed_2/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json +14 -0
  29. Blood/seed_2/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/model.safetensors.index.json +443 -0
  30. Blood/seed_2/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/special_tokens_map.json +31 -0
  31. Blood/seed_2/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/tokenizer_config.json +208 -0
  32. Blood/seed_4/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/README.md +67 -0
  33. Blood/seed_4/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/added_tokens.json +24 -0
  34. Blood/seed_4/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/chat_template.jinja +54 -0
  35. Blood/seed_4/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/config.json +66 -0
  36. Blood/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/added_tokens.json +24 -0
  37. Blood/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/chat_template.jinja +54 -0
  38. Blood/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/config.json +66 -0
  39. Blood/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json +14 -0
  40. Blood/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/added_tokens.json +24 -0
  41. Blood/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/chat_template.jinja +54 -0
  42. Blood/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/config.json +66 -0
  43. Blood/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json +14 -0
  44. Creditg/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/added_tokens.json +24 -0
  45. Creditg/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/chat_template.jinja +54 -0
  46. Creditg/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/config.json +66 -0
  47. Creditg/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/generation_config.json +14 -0
  48. Creditg/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/runs/Dec26_23-58-50_csce-yang-s2.engr.tamu.edu/events.out.tfevents.1766815143.csce-yang-s2.engr.tamu.edu.3648358.0 +3 -0
  49. Creditg/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/runs/Dec27_07-28-09_csce-yang-s2.engr.tamu.edu/events.out.tfevents.1766842103.csce-yang-s2.engr.tamu.edu.3873179.0 +3 -0
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ model_name: LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized
4
+ tags:
5
+ - generated_from_trainer
6
+ - grpo
7
+ - trl
8
+ licence: license
9
+ ---
10
+
11
+ # Model Card for LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized
12
+
13
+ This model is a fine-tuned version of [None](https://huggingface.co/None).
14
+ It has been trained using [TRL](https://github.com/huggingface/trl).
15
+
16
+ ## Quick start
17
+
18
+ ```python
19
+ from transformers import pipeline
20
+
21
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
22
+ generator = pipeline("text-generation", model="None", device="cuda")
23
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
24
+ print(output["generated_text"])
25
+ ```
26
+
27
+ ## Training procedure
28
+
29
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/baileyyeah326/MLLM/runs/hkmta5lx)
30
+
31
+
32
+ This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
33
+
34
+ ### Framework versions
35
+
36
+ - TRL: 0.21.0
37
+ - Transformers: 4.55.2
38
+ - Pytorch: 2.4.0+cu124
39
+ - Datasets: 4.0.0
40
+ - Tokenizers: 0.21.4
41
+
42
+ ## Citations
43
+
44
+ Cite GRPO as:
45
+
46
+ ```bibtex
47
+ @article{zhihong2024deepseekmath,
48
+ title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
49
+ author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
50
+ year = 2024,
51
+ eprint = {arXiv:2402.03300},
52
+ }
53
+
54
+ ```
55
+
56
+ Cite TRL as:
57
+
58
+ ```bibtex
59
+ @misc{vonwerra2022trl,
60
+ title = {{TRL: Transformer Reinforcement Learning}},
61
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
62
+ year = 2020,
63
+ journal = {GitHub repository},
64
+ publisher = {GitHub},
65
+ howpublished = {\url{https://github.com/huggingface/trl}}
66
+ }
67
+ ```
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 11008,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention"
49
+ ],
50
+ "max_position_embeddings": 32768,
51
+ "max_window_layers": 70,
52
+ "model_type": "qwen2",
53
+ "num_attention_heads": 16,
54
+ "num_hidden_layers": 36,
55
+ "num_key_value_heads": 2,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.55.2",
63
+ "use_cache": false,
64
+ "use_sliding_window": false,
65
+ "vocab_size": 151936
66
+ }
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/model.safetensors.index.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_parameters": 3085938688,
4
+ "total_size": 6794207232
5
+ },
6
+ "weight_map": {
7
+ "lm_head.weight": "model-00002-of-00002.safetensors",
8
+ "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
9
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
10
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
11
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
12
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
13
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
14
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
15
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
16
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
17
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
18
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
19
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
20
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
21
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors",
22
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
23
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
24
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
25
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
26
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
27
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
28
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
29
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
30
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
31
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
32
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
33
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors",
34
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
35
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
36
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
37
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
38
+ "model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
39
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
40
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
41
+ "model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
42
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
43
+ "model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
44
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
45
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors",
46
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
47
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
48
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
49
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
50
+ "model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
51
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
52
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
53
+ "model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
54
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
55
+ "model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
56
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
57
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors",
58
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
59
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
60
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
61
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
62
+ "model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
63
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
64
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
65
+ "model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
66
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
67
+ "model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
68
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
69
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
70
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
71
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
72
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
73
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
74
+ "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
75
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
76
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
77
+ "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
78
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
79
+ "model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
80
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
81
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
82
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
83
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
84
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
85
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
86
+ "model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
87
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
88
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
89
+ "model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
90
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
91
+ "model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
92
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
93
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors",
94
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
95
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
96
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
97
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
98
+ "model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
99
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
100
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
101
+ "model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
102
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
103
+ "model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
104
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
105
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors",
106
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
107
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
108
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
109
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
110
+ "model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
111
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
112
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
113
+ "model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
114
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
115
+ "model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
116
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
117
+ "model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors",
118
+ "model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
119
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
120
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
121
+ "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
122
+ "model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
123
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
124
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
125
+ "model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
126
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
127
+ "model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
128
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
129
+ "model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
130
+ "model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
131
+ "model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
132
+ "model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
133
+ "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
134
+ "model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
135
+ "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
136
+ "model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
137
+ "model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
138
+ "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
139
+ "model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
140
+ "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
141
+ "model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
142
+ "model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
143
+ "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
144
+ "model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
145
+ "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
146
+ "model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
147
+ "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
148
+ "model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
149
+ "model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
150
+ "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
151
+ "model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
152
+ "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
153
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
154
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
155
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
156
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
157
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
158
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
159
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
160
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
161
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
162
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
163
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
164
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
165
+ "model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
166
+ "model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
167
+ "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
168
+ "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
169
+ "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
170
+ "model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
171
+ "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
172
+ "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
173
+ "model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
174
+ "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
175
+ "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
176
+ "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
177
+ "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
178
+ "model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
179
+ "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
180
+ "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
181
+ "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
182
+ "model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
183
+ "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
184
+ "model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
185
+ "model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
186
+ "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
187
+ "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
188
+ "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
189
+ "model.layers.22.input_layernorm.weight": "model-00001-of-00002.safetensors",
190
+ "model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
191
+ "model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
192
+ "model.layers.22.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
193
+ "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
194
+ "model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
195
+ "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
196
+ "model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
197
+ "model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
198
+ "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
199
+ "model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
200
+ "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
201
+ "model.layers.23.input_layernorm.weight": "model-00001-of-00002.safetensors",
202
+ "model.layers.23.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
203
+ "model.layers.23.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
204
+ "model.layers.23.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
205
+ "model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
206
+ "model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
207
+ "model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
208
+ "model.layers.23.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
209
+ "model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
210
+ "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
211
+ "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
212
+ "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
213
+ "model.layers.24.input_layernorm.weight": "model-00001-of-00002.safetensors",
214
+ "model.layers.24.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
215
+ "model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
216
+ "model.layers.24.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
217
+ "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
218
+ "model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
219
+ "model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
220
+ "model.layers.24.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
221
+ "model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
222
+ "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
223
+ "model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
224
+ "model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
225
+ "model.layers.25.input_layernorm.weight": "model-00001-of-00002.safetensors",
226
+ "model.layers.25.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
227
+ "model.layers.25.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
228
+ "model.layers.25.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
229
+ "model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
230
+ "model.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
231
+ "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
232
+ "model.layers.25.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
233
+ "model.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
234
+ "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
235
+ "model.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
236
+ "model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
237
+ "model.layers.26.input_layernorm.weight": "model-00001-of-00002.safetensors",
238
+ "model.layers.26.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
239
+ "model.layers.26.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
240
+ "model.layers.26.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
241
+ "model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
242
+ "model.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
243
+ "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
244
+ "model.layers.26.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
245
+ "model.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
246
+ "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
247
+ "model.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
248
+ "model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
249
+ "model.layers.27.input_layernorm.weight": "model-00001-of-00002.safetensors",
250
+ "model.layers.27.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
251
+ "model.layers.27.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
252
+ "model.layers.27.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
253
+ "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
254
+ "model.layers.27.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
255
+ "model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
256
+ "model.layers.27.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
257
+ "model.layers.27.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
258
+ "model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
259
+ "model.layers.27.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
260
+ "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
261
+ "model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors",
262
+ "model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
263
+ "model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
264
+ "model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
265
+ "model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
266
+ "model.layers.28.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
267
+ "model.layers.28.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
268
+ "model.layers.28.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
269
+ "model.layers.28.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
270
+ "model.layers.28.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
271
+ "model.layers.28.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
272
+ "model.layers.28.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
273
+ "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
274
+ "model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
275
+ "model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
276
+ "model.layers.29.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
277
+ "model.layers.29.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
278
+ "model.layers.29.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
279
+ "model.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
280
+ "model.layers.29.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
281
+ "model.layers.29.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
282
+ "model.layers.29.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
283
+ "model.layers.29.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
284
+ "model.layers.29.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
285
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
286
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
287
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
288
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
289
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
290
+ "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
291
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
292
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
293
+ "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
294
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
295
+ "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
296
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
297
+ "model.layers.30.input_layernorm.weight": "model-00002-of-00002.safetensors",
298
+ "model.layers.30.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
299
+ "model.layers.30.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
300
+ "model.layers.30.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
301
+ "model.layers.30.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
302
+ "model.layers.30.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
303
+ "model.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
304
+ "model.layers.30.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
305
+ "model.layers.30.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
306
+ "model.layers.30.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
307
+ "model.layers.30.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
308
+ "model.layers.30.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
309
+ "model.layers.31.input_layernorm.weight": "model-00002-of-00002.safetensors",
310
+ "model.layers.31.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
311
+ "model.layers.31.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
312
+ "model.layers.31.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
313
+ "model.layers.31.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
314
+ "model.layers.31.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
315
+ "model.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
316
+ "model.layers.31.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
317
+ "model.layers.31.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
318
+ "model.layers.31.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
319
+ "model.layers.31.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
320
+ "model.layers.31.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
321
+ "model.layers.32.input_layernorm.weight": "model-00002-of-00002.safetensors",
322
+ "model.layers.32.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
323
+ "model.layers.32.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
324
+ "model.layers.32.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
325
+ "model.layers.32.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
326
+ "model.layers.32.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
327
+ "model.layers.32.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
328
+ "model.layers.32.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
329
+ "model.layers.32.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
330
+ "model.layers.32.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
331
+ "model.layers.32.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
332
+ "model.layers.32.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
333
+ "model.layers.33.input_layernorm.weight": "model-00002-of-00002.safetensors",
334
+ "model.layers.33.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
335
+ "model.layers.33.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
336
+ "model.layers.33.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
337
+ "model.layers.33.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
338
+ "model.layers.33.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
339
+ "model.layers.33.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
340
+ "model.layers.33.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
341
+ "model.layers.33.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
342
+ "model.layers.33.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
343
+ "model.layers.33.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
344
+ "model.layers.33.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
345
+ "model.layers.34.input_layernorm.weight": "model-00002-of-00002.safetensors",
346
+ "model.layers.34.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
347
+ "model.layers.34.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
348
+ "model.layers.34.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
349
+ "model.layers.34.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
350
+ "model.layers.34.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
351
+ "model.layers.34.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
352
+ "model.layers.34.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
353
+ "model.layers.34.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
354
+ "model.layers.34.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
355
+ "model.layers.34.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
356
+ "model.layers.34.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
357
+ "model.layers.35.input_layernorm.weight": "model-00002-of-00002.safetensors",
358
+ "model.layers.35.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
359
+ "model.layers.35.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
360
+ "model.layers.35.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
361
+ "model.layers.35.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
362
+ "model.layers.35.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
363
+ "model.layers.35.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
364
+ "model.layers.35.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
365
+ "model.layers.35.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
366
+ "model.layers.35.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
367
+ "model.layers.35.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
368
+ "model.layers.35.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
369
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
370
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
371
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
372
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
373
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
374
+ "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
375
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
376
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
377
+ "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
378
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
379
+ "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
380
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
381
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
382
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
383
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
384
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
385
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
386
+ "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
387
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
388
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
389
+ "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
390
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
391
+ "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
392
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
393
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
394
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
395
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
396
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
397
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
398
+ "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
399
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
400
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
401
+ "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
402
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
403
+ "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
404
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
405
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors",
406
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
407
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
408
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
409
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
410
+ "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
411
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
412
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
413
+ "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
414
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
415
+ "model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
416
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
417
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors",
418
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
419
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
420
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
421
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
422
+ "model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
423
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
424
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
425
+ "model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
426
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
427
+ "model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
428
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
429
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors",
430
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
431
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
432
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
433
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
434
+ "model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
435
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
436
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
437
+ "model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
438
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
439
+ "model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
440
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
441
+ "model.norm.weight": "model-00002-of-00002.safetensors"
442
+ }
443
+ }
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/tokenizer_config.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
+ "errors": "replace",
201
+ "extra_special_tokens": {},
202
+ "model_max_length": 2048,
203
+ "pad_token": "<|endoftext|>",
204
+ "padding_side": "right",
205
+ "split_special_tokens": false,
206
+ "tokenizer_class": "Qwen2Tokenizer",
207
+ "unk_token": null
208
+ }
Blood/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/trainer_state.json ADDED
@@ -0,0 +1,2331 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 400,
3
+ "best_metric": 0.7866666666666666,
4
+ "best_model_checkpoint": "/data/chenlang/Med3DVLM/llm/Blood/seeds/new/seed_0/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/checkpoint-400",
5
+ "epoch": 3.0,
6
+ "eval_steps": 100,
7
+ "global_step": 3588,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval_completions/clipped_ratio": 0.0,
15
+ "eval_completions/max_length": 490.5,
16
+ "eval_completions/max_terminated_length": 490.5,
17
+ "eval_completions/mean_length": 417.03125,
18
+ "eval_completions/mean_terminated_length": 417.03125,
19
+ "eval_completions/min_length": 321.0,
20
+ "eval_completions/min_terminated_length": 321.0,
21
+ "eval_frac_reward_zero_std": 0.0,
22
+ "eval_loss": 0.0,
23
+ "eval_num_tokens": 0.0,
24
+ "eval_reward": 0.7265625,
25
+ "eval_reward_std": 0.0,
26
+ "eval_rewards/accuracy_reward/mean": 0.7265625,
27
+ "eval_rewards/accuracy_reward/std": 0.44706645607948303,
28
+ "eval_runtime": 56.9822,
29
+ "eval_samples_per_second": 1.316,
30
+ "eval_steps_per_second": 0.035,
31
+ "step": 0
32
+ },
33
+ {
34
+ "clip_ratio/high_max": 0.0007908888192105223,
35
+ "clip_ratio/high_mean": 0.00039544440960526117,
36
+ "clip_ratio/low_mean": 4.6173167902452406e-06,
37
+ "clip_ratio/low_min": 0.0,
38
+ "clip_ratio/region_mean": 0.0004000617263955064,
39
+ "completions/clipped_ratio": 0.0,
40
+ "completions/max_length": 476.9375,
41
+ "completions/max_terminated_length": 476.9375,
42
+ "completions/mean_length": 415.46875,
43
+ "completions/mean_terminated_length": 415.46875,
44
+ "completions/min_length": 356.75,
45
+ "completions/min_terminated_length": 356.75,
46
+ "entropy": 0.09742542379535735,
47
+ "epoch": 0.05351170568561873,
48
+ "frac_reward_zero_std": 0.84375,
49
+ "grad_norm": 0.07067517191171646,
50
+ "kl": 0.0019223751488652852,
51
+ "learning_rate": 1.9648829431438127e-06,
52
+ "loss": 0.0004,
53
+ "num_tokens": 346848.0,
54
+ "reward": 0.849609375,
55
+ "reward_std": 0.052045637741684914,
56
+ "rewards/accuracy_reward/mean": 0.849609375,
57
+ "rewards/accuracy_reward/std": 0.18268440011888742,
58
+ "step": 64
59
+ },
60
+ {
61
+ "epoch": 0.08361204013377926,
62
+ "eval_completions/clipped_ratio": 0.0,
63
+ "eval_completions/max_length": 509.0,
64
+ "eval_completions/max_terminated_length": 509.0,
65
+ "eval_completions/mean_length": 416.984375,
66
+ "eval_completions/mean_terminated_length": 416.984375,
67
+ "eval_completions/min_length": 327.5,
68
+ "eval_completions/min_terminated_length": 327.5,
69
+ "eval_frac_reward_zero_std": 0.0,
70
+ "eval_loss": 0.0,
71
+ "eval_num_tokens": 541184.0,
72
+ "eval_reward": 0.7578125,
73
+ "eval_reward_std": 0.0,
74
+ "eval_rewards/accuracy_reward/mean": 0.7578125,
75
+ "eval_rewards/accuracy_reward/std": 0.4309118092060089,
76
+ "eval_runtime": 58.9151,
77
+ "eval_samples_per_second": 1.273,
78
+ "eval_steps_per_second": 0.034,
79
+ "step": 100
80
+ },
81
+ {
82
+ "clip_ratio/high_max": 0.0003593869956343302,
83
+ "clip_ratio/high_mean": 0.0003047454187513462,
84
+ "clip_ratio/low_mean": 0.0001279211137443781,
85
+ "clip_ratio/low_min": 4.644250397437385e-05,
86
+ "clip_ratio/region_mean": 0.0004326665324957243,
87
+ "completions/clipped_ratio": 0.0,
88
+ "completions/max_length": 470.57142857142856,
89
+ "completions/max_terminated_length": 470.57142857142856,
90
+ "completions/mean_length": 413.88392857142856,
91
+ "completions/mean_terminated_length": 413.88392857142856,
92
+ "completions/min_length": 325.0,
93
+ "completions/min_terminated_length": 325.0,
94
+ "entropy": 0.09871216783566135,
95
+ "epoch": 0.10702341137123746,
96
+ "frac_reward_zero_std": 0.8571428571428571,
97
+ "grad_norm": 0.08249606937170029,
98
+ "kl": 0.006351065875995638,
99
+ "learning_rate": 1.9292084726867336e-06,
100
+ "loss": -0.0037,
101
+ "num_tokens": 692662.0,
102
+ "reward": 0.9866071428571429,
103
+ "reward_std": 0.04225464378084455,
104
+ "rewards/accuracy_reward/mean": 0.9866071428571429,
105
+ "rewards/accuracy_reward/std": 0.042306367840085714,
106
+ "step": 128
107
+ },
108
+ {
109
+ "clip_ratio/high_max": 0.00019411579887673724,
110
+ "clip_ratio/high_mean": 9.705789943836862e-05,
111
+ "clip_ratio/low_mean": 0.0002299350276189216,
112
+ "clip_ratio/low_min": 0.0,
113
+ "clip_ratio/region_mean": 0.00032699292478355346,
114
+ "completions/clipped_ratio": 0.0,
115
+ "completions/max_length": 477.125,
116
+ "completions/max_terminated_length": 477.125,
117
+ "completions/mean_length": 427.84765625,
118
+ "completions/mean_terminated_length": 427.84765625,
119
+ "completions/min_length": 375.1875,
120
+ "completions/min_terminated_length": 375.1875,
121
+ "entropy": 0.09535896719899029,
122
+ "epoch": 0.1605351170568562,
123
+ "frac_reward_zero_std": 0.90625,
124
+ "grad_norm": 2.7755444049835205,
125
+ "kl": 0.008481005030262168,
126
+ "learning_rate": 1.8935340022296544e-06,
127
+ "loss": 0.0005,
128
+ "num_tokens": 1045752.0,
129
+ "reward": 0.845703125,
130
+ "reward_std": 0.0234375,
131
+ "rewards/accuracy_reward/mean": 0.845703125,
132
+ "rewards/accuracy_reward/std": 0.16967455577105284,
133
+ "step": 192
134
+ },
135
+ {
136
+ "epoch": 0.16722408026755853,
137
+ "eval_completions/clipped_ratio": 0.0,
138
+ "eval_completions/max_length": 507.0,
139
+ "eval_completions/max_terminated_length": 507.0,
140
+ "eval_completions/mean_length": 411.984375,
141
+ "eval_completions/mean_terminated_length": 411.984375,
142
+ "eval_completions/min_length": 329.0,
143
+ "eval_completions/min_terminated_length": 329.0,
144
+ "eval_frac_reward_zero_std": 0.0,
145
+ "eval_loss": 0.0,
146
+ "eval_num_tokens": 1088259.0,
147
+ "eval_reward": 0.7265625,
148
+ "eval_reward_std": 0.0,
149
+ "eval_rewards/accuracy_reward/mean": 0.7265625,
150
+ "eval_rewards/accuracy_reward/std": 0.4484642744064331,
151
+ "eval_runtime": 59.3904,
152
+ "eval_samples_per_second": 1.263,
153
+ "eval_steps_per_second": 0.034,
154
+ "step": 200
155
+ },
156
+ {
157
+ "clip_ratio/high_max": 0.0006860934559621715,
158
+ "clip_ratio/high_mean": 0.00034304672798108574,
159
+ "clip_ratio/low_mean": 0.0001461557621951215,
160
+ "clip_ratio/low_min": 0.0,
161
+ "clip_ratio/region_mean": 0.0004892024849790946,
162
+ "completions/clipped_ratio": 0.0,
163
+ "completions/max_length": 461.85714285714283,
164
+ "completions/max_terminated_length": 461.85714285714283,
165
+ "completions/mean_length": 405.6763392857143,
166
+ "completions/mean_terminated_length": 405.6763392857143,
167
+ "completions/min_length": 341.07142857142856,
168
+ "completions/min_terminated_length": 341.07142857142856,
169
+ "entropy": 0.09618141941194024,
170
+ "epoch": 0.2140468227424749,
171
+ "frac_reward_zero_std": 0.8928571428571429,
172
+ "grad_norm": 0.05106227472424507,
173
+ "kl": 0.015624415898595803,
174
+ "learning_rate": 1.8578595317725752e-06,
175
+ "loss": 0.0031,
176
+ "num_tokens": 1387378.0,
177
+ "reward": 0.8973214285714286,
178
+ "reward_std": 0.046613442046301704,
179
+ "rewards/accuracy_reward/mean": 0.8973214285714286,
180
+ "rewards/accuracy_reward/std": 0.14556341192552022,
181
+ "step": 256
182
+ },
183
+ {
184
+ "epoch": 0.2508361204013378,
185
+ "eval_completions/clipped_ratio": 0.0,
186
+ "eval_completions/max_length": 497.5,
187
+ "eval_completions/max_terminated_length": 497.5,
188
+ "eval_completions/mean_length": 413.3828125,
189
+ "eval_completions/mean_terminated_length": 413.3828125,
190
+ "eval_completions/min_length": 317.5,
191
+ "eval_completions/min_terminated_length": 317.5,
192
+ "eval_frac_reward_zero_std": 0.0,
193
+ "eval_loss": 0.0,
194
+ "eval_num_tokens": 1628714.0,
195
+ "eval_reward": 0.75,
196
+ "eval_reward_std": 0.0,
197
+ "eval_rewards/accuracy_reward/mean": 0.75,
198
+ "eval_rewards/accuracy_reward/std": 0.43605661392211914,
199
+ "eval_runtime": 57.5325,
200
+ "eval_samples_per_second": 1.304,
201
+ "eval_steps_per_second": 0.035,
202
+ "step": 300
203
+ },
204
+ {
205
+ "clip_ratio/high_max": 0.00016558784700464456,
206
+ "clip_ratio/high_mean": 8.279392350232228e-05,
207
+ "clip_ratio/low_mean": 0.0005344158911611885,
208
+ "clip_ratio/low_min": 0.0,
209
+ "clip_ratio/region_mean": 0.0006172098161187023,
210
+ "completions/clipped_ratio": 0.0,
211
+ "completions/max_length": 447.6,
212
+ "completions/max_terminated_length": 447.6,
213
+ "completions/mean_length": 397.49375,
214
+ "completions/mean_terminated_length": 397.49375,
215
+ "completions/min_length": 365.2,
216
+ "completions/min_terminated_length": 365.2,
217
+ "entropy": 0.08891196176409721,
218
+ "epoch": 0.26755852842809363,
219
+ "frac_reward_zero_std": 0.9,
220
+ "grad_norm": 0.10075017064809799,
221
+ "kl": 0.019166122027672827,
222
+ "learning_rate": 1.822185061315496e-06,
223
+ "loss": 0.0015,
224
+ "num_tokens": 1734249.0,
225
+ "reward": 0.425,
226
+ "reward_std": 0.04472135901451111,
227
+ "rewards/accuracy_reward/mean": 0.425,
228
+ "rewards/accuracy_reward/std": 0.1999739706516266,
229
+ "step": 320
230
+ },
231
+ {
232
+ "clip_ratio/high_max": 0.0003638084035628708,
233
+ "clip_ratio/high_mean": 0.0001819042017814354,
234
+ "clip_ratio/low_mean": 0.0002965977841995482,
235
+ "clip_ratio/low_min": 0.0,
236
+ "clip_ratio/region_mean": 0.00047850198188825743,
237
+ "completions/clipped_ratio": 0.0,
238
+ "completions/max_length": 468.75,
239
+ "completions/max_terminated_length": 468.75,
240
+ "completions/mean_length": 408.599609375,
241
+ "completions/mean_terminated_length": 408.599609375,
242
+ "completions/min_length": 358.375,
243
+ "completions/min_terminated_length": 358.375,
244
+ "entropy": 0.09487172344233841,
245
+ "epoch": 0.3210702341137124,
246
+ "frac_reward_zero_std": 0.875,
247
+ "grad_norm": 1.0121594667434692,
248
+ "kl": 0.021883822595555102,
249
+ "learning_rate": 1.786510590858417e-06,
250
+ "loss": 0.0023,
251
+ "num_tokens": 2077548.0,
252
+ "reward": 0.67578125,
253
+ "reward_std": 0.044233137741684914,
254
+ "rewards/accuracy_reward/mean": 0.67578125,
255
+ "rewards/accuracy_reward/std": 0.3040844462811947,
256
+ "step": 384
257
+ },
258
+ {
259
+ "epoch": 0.33444816053511706,
260
+ "eval_completions/clipped_ratio": 0.0,
261
+ "eval_completions/max_length": 493.0,
262
+ "eval_completions/max_terminated_length": 493.0,
263
+ "eval_completions/mean_length": 417.265625,
264
+ "eval_completions/mean_terminated_length": 417.265625,
265
+ "eval_completions/min_length": 321.5,
266
+ "eval_completions/min_terminated_length": 321.5,
267
+ "eval_frac_reward_zero_std": 0.0,
268
+ "eval_loss": 0.0,
269
+ "eval_num_tokens": 2168096.0,
270
+ "eval_reward": 0.75,
271
+ "eval_reward_std": 0.0,
272
+ "eval_rewards/accuracy_reward/mean": 0.75,
273
+ "eval_rewards/accuracy_reward/std": 0.4364357888698578,
274
+ "eval_runtime": 57.8802,
275
+ "eval_samples_per_second": 1.296,
276
+ "eval_steps_per_second": 0.035,
277
+ "step": 400
278
+ },
279
+ {
280
+ "clip_ratio/high_max": 0.0008308086750427416,
281
+ "clip_ratio/high_mean": 0.000467669527400479,
282
+ "clip_ratio/low_mean": 0.0007246423750378502,
283
+ "clip_ratio/low_min": 0.00011243560341730093,
284
+ "clip_ratio/region_mean": 0.0011923118945560418,
285
+ "completions/clipped_ratio": 0.0,
286
+ "completions/max_length": 472.5,
287
+ "completions/max_terminated_length": 472.5,
288
+ "completions/mean_length": 418.3307291666667,
289
+ "completions/mean_terminated_length": 418.3307291666667,
290
+ "completions/min_length": 360.8333333333333,
291
+ "completions/min_terminated_length": 360.8333333333333,
292
+ "entropy": 0.1242803984011213,
293
+ "epoch": 0.3745819397993311,
294
+ "frac_reward_zero_std": 0.7083333333333334,
295
+ "grad_norm": 0.8793007135391235,
296
+ "kl": 0.07019452709937468,
297
+ "learning_rate": 1.7508361204013377e-06,
298
+ "loss": 0.0067,
299
+ "num_tokens": 2429423.0,
300
+ "reward": 0.734375,
301
+ "reward_std": 0.11149696509043376,
302
+ "rewards/accuracy_reward/mean": 0.734375,
303
+ "rewards/accuracy_reward/std": 0.25355326011776924,
304
+ "step": 448
305
+ },
306
+ {
307
+ "epoch": 0.4180602006688963,
308
+ "eval_completions/clipped_ratio": 0.0,
309
+ "eval_completions/max_length": 495.5,
310
+ "eval_completions/max_terminated_length": 495.5,
311
+ "eval_completions/mean_length": 413.9140625,
312
+ "eval_completions/mean_terminated_length": 413.9140625,
313
+ "eval_completions/min_length": 318.5,
314
+ "eval_completions/min_terminated_length": 318.5,
315
+ "eval_frac_reward_zero_std": 0.0,
316
+ "eval_loss": 0.0,
317
+ "eval_num_tokens": 2717672.0,
318
+ "eval_reward": 0.734375,
319
+ "eval_reward_std": 0.0,
320
+ "eval_rewards/accuracy_reward/mean": 0.734375,
321
+ "eval_rewards/accuracy_reward/std": 0.44479964673519135,
322
+ "eval_runtime": 58.746,
323
+ "eval_samples_per_second": 1.277,
324
+ "eval_steps_per_second": 0.034,
325
+ "step": 500
326
+ },
327
+ {
328
+ "clip_ratio/high_max": 0.000866840787542363,
329
+ "clip_ratio/high_mean": 0.0004334203937711815,
330
+ "clip_ratio/low_mean": 0.00020594061546338102,
331
+ "clip_ratio/low_min": 0.0,
332
+ "clip_ratio/region_mean": 0.0006393609849813705,
333
+ "completions/clipped_ratio": 0.0,
334
+ "completions/max_length": 479.3333333333333,
335
+ "completions/max_terminated_length": 479.3333333333333,
336
+ "completions/mean_length": 420.9791666666667,
337
+ "completions/mean_terminated_length": 420.9791666666667,
338
+ "completions/min_length": 365.6666666666667,
339
+ "completions/min_terminated_length": 365.6666666666667,
340
+ "entropy": 0.12591193988919258,
341
+ "epoch": 0.4280936454849498,
342
+ "frac_reward_zero_std": 0.8333333333333334,
343
+ "grad_norm": 0.09410286694765091,
344
+ "kl": 0.05266569306453069,
345
+ "learning_rate": 1.7151616499442586e-06,
346
+ "loss": 0.0056,
347
+ "num_tokens": 2783286.0,
348
+ "reward": 0.96875,
349
+ "reward_std": 0.0671854813893636,
350
+ "rewards/accuracy_reward/mean": 0.96875,
351
+ "rewards/accuracy_reward/std": 0.09871485829353333,
352
+ "step": 512
353
+ },
354
+ {
355
+ "clip_ratio/high_max": 7.435129555233289e-05,
356
+ "clip_ratio/high_mean": 3.717564777616644e-05,
357
+ "clip_ratio/low_mean": 0.0,
358
+ "clip_ratio/low_min": 0.0,
359
+ "clip_ratio/region_mean": 3.717564777616644e-05,
360
+ "completions/clipped_ratio": 0.0,
361
+ "completions/max_length": 475.8125,
362
+ "completions/max_terminated_length": 475.8125,
363
+ "completions/mean_length": 421.255859375,
364
+ "completions/mean_terminated_length": 421.255859375,
365
+ "completions/min_length": 365.25,
366
+ "completions/min_terminated_length": 365.25,
367
+ "entropy": 0.11762141226790845,
368
+ "epoch": 0.4816053511705686,
369
+ "frac_reward_zero_std": 0.96875,
370
+ "grad_norm": 0.08674240112304688,
371
+ "kl": 0.030779874985455535,
372
+ "learning_rate": 1.6794871794871794e-06,
373
+ "loss": 0.003,
374
+ "num_tokens": 3133097.0,
375
+ "reward": 0.80859375,
376
+ "reward_std": 0.010673906654119492,
377
+ "rewards/accuracy_reward/mean": 0.80859375,
378
+ "rewards/accuracy_reward/std": 0.06325103715062141,
379
+ "step": 576
380
+ },
381
+ {
382
+ "epoch": 0.5016722408026756,
383
+ "eval_completions/clipped_ratio": 0.0,
384
+ "eval_completions/max_length": 491.5,
385
+ "eval_completions/max_terminated_length": 491.5,
386
+ "eval_completions/mean_length": 415.6484375,
387
+ "eval_completions/mean_terminated_length": 415.6484375,
388
+ "eval_completions/min_length": 318.0,
389
+ "eval_completions/min_terminated_length": 318.0,
390
+ "eval_frac_reward_zero_std": 0.0,
391
+ "eval_loss": 0.0,
392
+ "eval_num_tokens": 3263273.0,
393
+ "eval_reward": 0.7421875,
394
+ "eval_reward_std": 0.0,
395
+ "eval_rewards/accuracy_reward/mean": 0.7421875,
396
+ "eval_rewards/accuracy_reward/std": 0.440796360373497,
397
+ "eval_runtime": 56.5774,
398
+ "eval_samples_per_second": 1.326,
399
+ "eval_steps_per_second": 0.035,
400
+ "step": 600
401
+ },
402
+ {
403
+ "clip_ratio/high_max": 4.976774798706174e-05,
404
+ "clip_ratio/high_mean": 2.488387399353087e-05,
405
+ "clip_ratio/low_mean": 0.00016385155322495848,
406
+ "clip_ratio/low_min": 0.0,
407
+ "clip_ratio/region_mean": 0.00018873543303925543,
408
+ "completions/clipped_ratio": 0.0,
409
+ "completions/max_length": 474.7,
410
+ "completions/max_terminated_length": 474.7,
411
+ "completions/mean_length": 411.803125,
412
+ "completions/mean_terminated_length": 411.803125,
413
+ "completions/min_length": 346.0,
414
+ "completions/min_terminated_length": 346.0,
415
+ "entropy": 0.11736641991883516,
416
+ "epoch": 0.5351170568561873,
417
+ "frac_reward_zero_std": 0.95,
418
+ "grad_norm": 1.3029202222824097,
419
+ "kl": 0.026357294467743486,
420
+ "learning_rate": 1.6438127090301002e-06,
421
+ "loss": 0.0028,
422
+ "num_tokens": 3478874.0,
423
+ "reward": 0.609375,
424
+ "reward_std": 0.02015564441680908,
425
+ "rewards/accuracy_reward/mean": 0.609375,
426
+ "rewards/accuracy_reward/std": 0.30389933586120604,
427
+ "step": 640
428
+ },
429
+ {
430
+ "epoch": 0.5852842809364549,
431
+ "eval_completions/clipped_ratio": 0.0,
432
+ "eval_completions/max_length": 491.0,
433
+ "eval_completions/max_terminated_length": 491.0,
434
+ "eval_completions/mean_length": 408.3203125,
435
+ "eval_completions/mean_terminated_length": 408.3203125,
436
+ "eval_completions/min_length": 305.0,
437
+ "eval_completions/min_terminated_length": 305.0,
438
+ "eval_frac_reward_zero_std": 0.0,
439
+ "eval_loss": 0.0,
440
+ "eval_num_tokens": 3800664.0,
441
+ "eval_reward": 0.703125,
442
+ "eval_reward_std": 0.0,
443
+ "eval_rewards/accuracy_reward/mean": 0.703125,
444
+ "eval_rewards/accuracy_reward/std": 0.4552833288908005,
445
+ "eval_runtime": 57.1263,
446
+ "eval_samples_per_second": 1.313,
447
+ "eval_steps_per_second": 0.035,
448
+ "step": 700
449
+ },
450
+ {
451
+ "clip_ratio/high_max": 0.001476035060477443,
452
+ "clip_ratio/high_mean": 0.0007380175302387215,
453
+ "clip_ratio/low_mean": 0.0,
454
+ "clip_ratio/low_min": 0.0,
455
+ "clip_ratio/region_mean": 0.0007380175302387215,
456
+ "completions/clipped_ratio": 0.0,
457
+ "completions/max_length": 507.0,
458
+ "completions/max_terminated_length": 507.0,
459
+ "completions/mean_length": 434.03125,
460
+ "completions/mean_terminated_length": 434.03125,
461
+ "completions/min_length": 369.0,
462
+ "completions/min_terminated_length": 369.0,
463
+ "entropy": 0.1387515664100647,
464
+ "epoch": 0.5886287625418061,
465
+ "frac_reward_zero_std": 0.5,
466
+ "grad_norm": 0.545124351978302,
467
+ "kl": 0.09541250206530094,
468
+ "learning_rate": 1.608138238573021e-06,
469
+ "loss": 0.0109,
470
+ "num_tokens": 3822969.0,
471
+ "reward": 0.96875,
472
+ "reward_std": 0.125,
473
+ "rewards/accuracy_reward/mean": 0.96875,
474
+ "rewards/accuracy_reward/std": 0.1767766922712326,
475
+ "step": 704
476
+ },
477
+ {
478
+ "clip_ratio/high_max": 0.00048521713688387536,
479
+ "clip_ratio/high_mean": 0.00024260856844193768,
480
+ "clip_ratio/low_mean": 7.099746426320053e-05,
481
+ "clip_ratio/low_min": 0.0,
482
+ "clip_ratio/region_mean": 0.00031360603679786436,
483
+ "completions/clipped_ratio": 0.0,
484
+ "completions/max_length": 469.3125,
485
+ "completions/max_terminated_length": 469.3125,
486
+ "completions/mean_length": 403.490234375,
487
+ "completions/mean_terminated_length": 403.490234375,
488
+ "completions/min_length": 345.0625,
489
+ "completions/min_terminated_length": 345.0625,
490
+ "entropy": 0.12115569307934493,
491
+ "epoch": 0.6421404682274248,
492
+ "frac_reward_zero_std": 0.875,
493
+ "grad_norm": 0.09112782031297684,
494
+ "kl": 0.03751233006914845,
495
+ "learning_rate": 1.5724637681159421e-06,
496
+ "loss": 0.0046,
497
+ "num_tokens": 4163780.0,
498
+ "reward": 0.8359375,
499
+ "reward_std": 0.05667105130851269,
500
+ "rewards/accuracy_reward/mean": 0.8359375,
501
+ "rewards/accuracy_reward/std": 0.21587339136749506,
502
+ "step": 768
503
+ },
504
+ {
505
+ "epoch": 0.6688963210702341,
506
+ "eval_completions/clipped_ratio": 0.0,
507
+ "eval_completions/max_length": 498.0,
508
+ "eval_completions/max_terminated_length": 498.0,
509
+ "eval_completions/mean_length": 418.78125,
510
+ "eval_completions/mean_terminated_length": 418.78125,
511
+ "eval_completions/min_length": 325.0,
512
+ "eval_completions/min_terminated_length": 325.0,
513
+ "eval_frac_reward_zero_std": 0.0,
514
+ "eval_loss": 0.0,
515
+ "eval_num_tokens": 4336339.0,
516
+ "eval_reward": 0.7109375,
517
+ "eval_reward_std": 0.0,
518
+ "eval_rewards/accuracy_reward/mean": 0.7109375,
519
+ "eval_rewards/accuracy_reward/std": 0.4561667740345001,
520
+ "eval_runtime": 58.6319,
521
+ "eval_samples_per_second": 1.279,
522
+ "eval_steps_per_second": 0.034,
523
+ "step": 800
524
+ },
525
+ {
526
+ "clip_ratio/high_max": 0.0,
527
+ "clip_ratio/high_mean": 0.0,
528
+ "clip_ratio/low_mean": 0.0,
529
+ "clip_ratio/low_min": 0.0,
530
+ "clip_ratio/region_mean": 0.0,
531
+ "completions/clipped_ratio": 0.0,
532
+ "completions/max_length": 457.25,
533
+ "completions/max_terminated_length": 457.25,
534
+ "completions/mean_length": 402.98046875,
535
+ "completions/mean_terminated_length": 402.98046875,
536
+ "completions/min_length": 344.0,
537
+ "completions/min_terminated_length": 344.0,
538
+ "entropy": 0.11502773710526526,
539
+ "epoch": 0.6956521739130435,
540
+ "frac_reward_zero_std": 1.0,
541
+ "grad_norm": 0.23805662989616394,
542
+ "kl": 0.01497033271152759,
543
+ "learning_rate": 1.5367892976588628e-06,
544
+ "loss": 0.0029,
545
+ "num_tokens": 4506622.0,
546
+ "reward": 0.875,
547
+ "reward_std": 0.0,
548
+ "rewards/accuracy_reward/mean": 0.875,
549
+ "rewards/accuracy_reward/std": 0.1270001232624054,
550
+ "step": 832
551
+ },
552
+ {
553
+ "clip_ratio/high_max": 0.0009068004947039299,
554
+ "clip_ratio/high_mean": 0.0004959487878295477,
555
+ "clip_ratio/low_mean": 5.776368061560788e-05,
556
+ "clip_ratio/low_min": 0.0,
557
+ "clip_ratio/region_mean": 0.000553712471628387,
558
+ "completions/clipped_ratio": 0.0,
559
+ "completions/max_length": 471.5625,
560
+ "completions/max_terminated_length": 471.5625,
561
+ "completions/mean_length": 414.232421875,
562
+ "completions/mean_terminated_length": 414.232421875,
563
+ "completions/min_length": 365.375,
564
+ "completions/min_terminated_length": 365.375,
565
+ "entropy": 0.129343748325482,
566
+ "epoch": 0.7491638795986622,
567
+ "frac_reward_zero_std": 0.75,
568
+ "grad_norm": 0.22267663478851318,
569
+ "kl": 0.06127101270976709,
570
+ "learning_rate": 1.5011148272017838e-06,
571
+ "loss": 0.0099,
572
+ "num_tokens": 4852805.0,
573
+ "reward": 0.873046875,
574
+ "reward_std": 0.08301629684865475,
575
+ "rewards/accuracy_reward/mean": 0.873046875,
576
+ "rewards/accuracy_reward/std": 0.19705284852534533,
577
+ "step": 896
578
+ },
579
+ {
580
+ "epoch": 0.7525083612040134,
581
+ "eval_completions/clipped_ratio": 0.0,
582
+ "eval_completions/max_length": 493.0,
583
+ "eval_completions/max_terminated_length": 493.0,
584
+ "eval_completions/mean_length": 413.3203125,
585
+ "eval_completions/mean_terminated_length": 413.3203125,
586
+ "eval_completions/min_length": 329.5,
587
+ "eval_completions/min_terminated_length": 329.5,
588
+ "eval_frac_reward_zero_std": 0.0,
589
+ "eval_loss": 0.0,
590
+ "eval_num_tokens": 4873883.0,
591
+ "eval_reward": 0.75,
592
+ "eval_reward_std": 0.0,
593
+ "eval_rewards/accuracy_reward/mean": 0.75,
594
+ "eval_rewards/accuracy_reward/std": 0.43491509556770325,
595
+ "eval_runtime": 56.8767,
596
+ "eval_samples_per_second": 1.319,
597
+ "eval_steps_per_second": 0.035,
598
+ "step": 900
599
+ },
600
+ {
601
+ "clip_ratio/high_max": 0.0004170371171009416,
602
+ "clip_ratio/high_mean": 0.0002085185585504708,
603
+ "clip_ratio/low_mean": 0.00011040495689182231,
604
+ "clip_ratio/low_min": 0.0,
605
+ "clip_ratio/region_mean": 0.00031892351592735697,
606
+ "completions/clipped_ratio": 0.0,
607
+ "completions/max_length": 466.3333333333333,
608
+ "completions/max_terminated_length": 466.3333333333333,
609
+ "completions/mean_length": 403.9375,
610
+ "completions/mean_terminated_length": 403.9375,
611
+ "completions/min_length": 349.8,
612
+ "completions/min_terminated_length": 349.8,
613
+ "entropy": 0.13929289107521375,
614
+ "epoch": 0.802675585284281,
615
+ "frac_reward_zero_std": 0.8333333333333334,
616
+ "grad_norm": 0.45232445001602173,
617
+ "kl": 0.04927637780395647,
618
+ "learning_rate": 1.4654403567447044e-06,
619
+ "loss": 0.0042,
620
+ "num_tokens": 5193645.0,
621
+ "reward": 0.8520833333333333,
622
+ "reward_std": 0.06744743982950846,
623
+ "rewards/accuracy_reward/mean": 0.8520833333333333,
624
+ "rewards/accuracy_reward/std": 0.20930681228637696,
625
+ "step": 960
626
+ },
627
+ {
628
+ "epoch": 0.8361204013377926,
629
+ "eval_completions/clipped_ratio": 0.0,
630
+ "eval_completions/max_length": 498.0,
631
+ "eval_completions/max_terminated_length": 498.0,
632
+ "eval_completions/mean_length": 414.2265625,
633
+ "eval_completions/mean_terminated_length": 414.2265625,
634
+ "eval_completions/min_length": 327.0,
635
+ "eval_completions/min_terminated_length": 327.0,
636
+ "eval_frac_reward_zero_std": 0.0,
637
+ "eval_loss": 0.0,
638
+ "eval_num_tokens": 5410791.0,
639
+ "eval_reward": 0.671875,
640
+ "eval_reward_std": 0.0,
641
+ "eval_rewards/accuracy_reward/mean": 0.671875,
642
+ "eval_rewards/accuracy_reward/std": 0.47205156087875366,
643
+ "eval_runtime": 57.8412,
644
+ "eval_samples_per_second": 1.297,
645
+ "eval_steps_per_second": 0.035,
646
+ "step": 1000
647
+ },
648
+ {
649
+ "clip_ratio/high_max": 0.0008678070929211875,
650
+ "clip_ratio/high_mean": 0.00043390354646059376,
651
+ "clip_ratio/low_mean": 0.0002630315866554156,
652
+ "clip_ratio/low_min": 0.0,
653
+ "clip_ratio/region_mean": 0.0006969351331160093,
654
+ "completions/clipped_ratio": 0.0,
655
+ "completions/max_length": 474.8333333333333,
656
+ "completions/max_terminated_length": 474.8333333333333,
657
+ "completions/mean_length": 412.7291666666667,
658
+ "completions/mean_terminated_length": 412.7291666666667,
659
+ "completions/min_length": 354.8333333333333,
660
+ "completions/min_terminated_length": 354.8333333333333,
661
+ "entropy": 0.14330726861953735,
662
+ "epoch": 0.8561872909698997,
663
+ "frac_reward_zero_std": 0.75,
664
+ "grad_norm": 1.0059927701950073,
665
+ "kl": 0.05156718186723689,
666
+ "learning_rate": 1.4297658862876255e-06,
667
+ "loss": 0.0074,
668
+ "num_tokens": 5540323.0,
669
+ "reward": 0.7135416666666666,
670
+ "reward_std": 0.10508963962395985,
671
+ "rewards/accuracy_reward/mean": 0.7135416666666666,
672
+ "rewards/accuracy_reward/std": 0.1636701002717018,
673
+ "step": 1024
674
+ },
675
+ {
676
+ "clip_ratio/high_max": 0.0006484732912213076,
677
+ "clip_ratio/high_mean": 0.0003492480500426609,
678
+ "clip_ratio/low_mean": 0.00020737830254802248,
679
+ "clip_ratio/low_min": 0.0,
680
+ "clip_ratio/region_mean": 0.0005566263575929042,
681
+ "completions/clipped_ratio": 0.0,
682
+ "completions/max_length": 478.6875,
683
+ "completions/max_terminated_length": 478.6875,
684
+ "completions/mean_length": 417.869140625,
685
+ "completions/mean_terminated_length": 417.869140625,
686
+ "completions/min_length": 353.6875,
687
+ "completions/min_terminated_length": 353.6875,
688
+ "entropy": 0.1362351783318445,
689
+ "epoch": 0.9096989966555183,
690
+ "frac_reward_zero_std": 0.6875,
691
+ "grad_norm": 1.5935475826263428,
692
+ "kl": 0.049660728640446905,
693
+ "learning_rate": 1.394091415830546e-06,
694
+ "loss": 0.0083,
695
+ "num_tokens": 5888384.0,
696
+ "reward": 0.80859375,
697
+ "reward_std": 0.10727976821362972,
698
+ "rewards/accuracy_reward/mean": 0.80859375,
699
+ "rewards/accuracy_reward/std": 0.24826626759022474,
700
+ "step": 1088
701
+ },
702
+ {
703
+ "epoch": 0.919732441471572,
704
+ "eval_completions/clipped_ratio": 0.0,
705
+ "eval_completions/max_length": 481.5,
706
+ "eval_completions/max_terminated_length": 481.5,
707
+ "eval_completions/mean_length": 414.1484375,
708
+ "eval_completions/mean_terminated_length": 414.1484375,
709
+ "eval_completions/min_length": 323.0,
710
+ "eval_completions/min_terminated_length": 323.0,
711
+ "eval_frac_reward_zero_std": 0.0,
712
+ "eval_loss": 0.0,
713
+ "eval_num_tokens": 5954092.0,
714
+ "eval_reward": 0.6796875,
715
+ "eval_reward_std": 0.0,
716
+ "eval_rewards/accuracy_reward/mean": 0.6796875,
717
+ "eval_rewards/accuracy_reward/std": 0.46960316598415375,
718
+ "eval_runtime": 55.3948,
719
+ "eval_samples_per_second": 1.354,
720
+ "eval_steps_per_second": 0.036,
721
+ "step": 1100
722
+ },
723
+ {
724
+ "clip_ratio/high_max": 0.00026838650228455663,
725
+ "clip_ratio/high_mean": 0.00013941049781771234,
726
+ "clip_ratio/low_mean": 0.0001278050331166014,
727
+ "clip_ratio/low_min": 0.0,
728
+ "clip_ratio/region_mean": 0.00026721552645680134,
729
+ "completions/clipped_ratio": 0.0,
730
+ "completions/max_length": 487.38461538461536,
731
+ "completions/max_terminated_length": 487.38461538461536,
732
+ "completions/mean_length": 433.93028846153845,
733
+ "completions/mean_terminated_length": 433.93028846153845,
734
+ "completions/min_length": 371.3076923076923,
735
+ "completions/min_terminated_length": 371.3076923076923,
736
+ "entropy": 0.2032681694970681,
737
+ "epoch": 0.9632107023411371,
738
+ "frac_reward_zero_std": 0.7307692307692307,
739
+ "grad_norm": 1.172003149986267,
740
+ "kl": 0.582390272954049,
741
+ "learning_rate": 1.3584169453734671e-06,
742
+ "loss": 0.0459,
743
+ "num_tokens": 6243567.0,
744
+ "reward": 0.8245192307692307,
745
+ "reward_std": 0.10265854688791129,
746
+ "rewards/accuracy_reward/mean": 0.8245192307692307,
747
+ "rewards/accuracy_reward/std": 0.1861229149194864,
748
+ "step": 1152
749
+ },
750
+ {
751
+ "epoch": 1.0033444816053512,
752
+ "eval_completions/clipped_ratio": 0.0,
753
+ "eval_completions/max_length": 484.0,
754
+ "eval_completions/max_terminated_length": 484.0,
755
+ "eval_completions/mean_length": 409.8359375,
756
+ "eval_completions/mean_terminated_length": 409.8359375,
757
+ "eval_completions/min_length": 290.5,
758
+ "eval_completions/min_terminated_length": 290.5,
759
+ "eval_frac_reward_zero_std": 0.0,
760
+ "eval_loss": 0.0,
761
+ "eval_num_tokens": 6498032.0,
762
+ "eval_reward": 0.7265625,
763
+ "eval_reward_std": 0.0,
764
+ "eval_rewards/accuracy_reward/mean": 0.7265625,
765
+ "eval_rewards/accuracy_reward/std": 0.4484642744064331,
766
+ "eval_runtime": 56.3146,
767
+ "eval_samples_per_second": 1.332,
768
+ "eval_steps_per_second": 0.036,
769
+ "step": 1200
770
+ },
771
+ {
772
+ "clip_ratio/high_max": 4.045307287015021e-05,
773
+ "clip_ratio/high_mean": 2.0226536435075104e-05,
774
+ "clip_ratio/low_mean": 0.0,
775
+ "clip_ratio/low_min": 0.0,
776
+ "clip_ratio/region_mean": 2.0226536435075104e-05,
777
+ "completions/clipped_ratio": 0.0,
778
+ "completions/max_length": 436.0,
779
+ "completions/max_terminated_length": 436.0,
780
+ "completions/mean_length": 365.421875,
781
+ "completions/mean_terminated_length": 365.421875,
782
+ "completions/min_length": 313.5,
783
+ "completions/min_terminated_length": 313.5,
784
+ "entropy": 0.2158376593142748,
785
+ "epoch": 1.0167224080267558,
786
+ "frac_reward_zero_std": 0.75,
787
+ "grad_norm": 0.1277727633714676,
788
+ "kl": 0.15024185786023736,
789
+ "learning_rate": 1.3227424749163878e-06,
790
+ "loss": 0.0152,
791
+ "num_tokens": 6578326.0,
792
+ "reward": 0.8828125,
793
+ "reward_std": 0.09308473765850067,
794
+ "rewards/accuracy_reward/mean": 0.8828125,
795
+ "rewards/accuracy_reward/std": 0.18623139336705208,
796
+ "step": 1216
797
+ },
798
+ {
799
+ "clip_ratio/high_max": 9.294958454120206e-05,
800
+ "clip_ratio/high_mean": 4.647479227060103e-05,
801
+ "clip_ratio/low_mean": 9.177984338748502e-06,
802
+ "clip_ratio/low_min": 0.0,
803
+ "clip_ratio/region_mean": 5.565277660934953e-05,
804
+ "completions/clipped_ratio": 0.0,
805
+ "completions/max_length": 478.5,
806
+ "completions/max_terminated_length": 478.5,
807
+ "completions/mean_length": 400.9921875,
808
+ "completions/mean_terminated_length": 400.9921875,
809
+ "completions/min_length": 326.5,
810
+ "completions/min_terminated_length": 326.5,
811
+ "entropy": 0.24563083425164223,
812
+ "epoch": 1.0702341137123745,
813
+ "frac_reward_zero_std": 0.625,
814
+ "grad_norm": 2.6599202156066895,
815
+ "kl": 0.1877364907413721,
816
+ "learning_rate": 1.2870680044593088e-06,
817
+ "loss": 0.0201,
818
+ "num_tokens": 6917794.0,
819
+ "reward": 0.849609375,
820
+ "reward_std": 0.12969218753278255,
821
+ "rewards/accuracy_reward/mean": 0.849609375,
822
+ "rewards/accuracy_reward/std": 0.2124346774071455,
823
+ "step": 1280
824
+ },
825
+ {
826
+ "epoch": 1.0869565217391304,
827
+ "eval_completions/clipped_ratio": 0.0,
828
+ "eval_completions/max_length": 497.0,
829
+ "eval_completions/max_terminated_length": 497.0,
830
+ "eval_completions/mean_length": 404.3125,
831
+ "eval_completions/mean_terminated_length": 404.3125,
832
+ "eval_completions/min_length": 283.5,
833
+ "eval_completions/min_terminated_length": 283.5,
834
+ "eval_frac_reward_zero_std": 0.0,
835
+ "eval_loss": 0.0,
836
+ "eval_num_tokens": 7025182.0,
837
+ "eval_reward": 0.6875,
838
+ "eval_reward_std": 0.0,
839
+ "eval_rewards/accuracy_reward/mean": 0.6875,
840
+ "eval_rewards/accuracy_reward/std": 0.467176616191864,
841
+ "eval_runtime": 58.4461,
842
+ "eval_samples_per_second": 1.283,
843
+ "eval_steps_per_second": 0.034,
844
+ "step": 1300
845
+ },
846
+ {
847
+ "clip_ratio/high_max": 8.207485503093763e-05,
848
+ "clip_ratio/high_mean": 4.1037427515468814e-05,
849
+ "clip_ratio/low_mean": 6.5723751586946575e-06,
850
+ "clip_ratio/low_min": 0.0,
851
+ "clip_ratio/region_mean": 4.760980267416347e-05,
852
+ "completions/clipped_ratio": 0.0,
853
+ "completions/max_length": 474.09090909090907,
854
+ "completions/max_terminated_length": 474.09090909090907,
855
+ "completions/mean_length": 394.9261363636364,
856
+ "completions/mean_terminated_length": 394.9261363636364,
857
+ "completions/min_length": 320.09090909090907,
858
+ "completions/min_terminated_length": 320.09090909090907,
859
+ "entropy": 0.2421540587463162,
860
+ "epoch": 1.1237458193979932,
861
+ "frac_reward_zero_std": 0.6363636363636364,
862
+ "grad_norm": 0.14775635302066803,
863
+ "kl": 0.1778960299085487,
864
+ "learning_rate": 1.2513935340022296e-06,
865
+ "loss": 0.0183,
866
+ "num_tokens": 7256404.0,
867
+ "reward": 0.71875,
868
+ "reward_std": 0.15349717302755875,
869
+ "rewards/accuracy_reward/mean": 0.71875,
870
+ "rewards/accuracy_reward/std": 0.3336510712450201,
871
+ "step": 1344
872
+ },
873
+ {
874
+ "epoch": 1.1705685618729098,
875
+ "eval_completions/clipped_ratio": 0.0,
876
+ "eval_completions/max_length": 512.0,
877
+ "eval_completions/max_terminated_length": 512.0,
878
+ "eval_completions/mean_length": 407.3515625,
879
+ "eval_completions/mean_terminated_length": 407.3515625,
880
+ "eval_completions/min_length": 305.5,
881
+ "eval_completions/min_terminated_length": 305.5,
882
+ "eval_frac_reward_zero_std": 0.0,
883
+ "eval_loss": 0.0,
884
+ "eval_num_tokens": 7560502.0,
885
+ "eval_reward": 0.71875,
886
+ "eval_reward_std": 0.0,
887
+ "eval_rewards/accuracy_reward/mean": 0.71875,
888
+ "eval_rewards/accuracy_reward/std": 0.4518062025308609,
889
+ "eval_runtime": 58.6946,
890
+ "eval_samples_per_second": 1.278,
891
+ "eval_steps_per_second": 0.034,
892
+ "step": 1400
893
+ },
894
+ {
895
+ "clip_ratio/high_max": 0.0,
896
+ "clip_ratio/high_mean": 0.0,
897
+ "clip_ratio/low_mean": 0.0,
898
+ "clip_ratio/low_min": 0.0,
899
+ "clip_ratio/region_mean": 0.0,
900
+ "completions/clipped_ratio": 0.0,
901
+ "completions/max_length": 469.5,
902
+ "completions/max_terminated_length": 469.5,
903
+ "completions/mean_length": 400.265625,
904
+ "completions/mean_terminated_length": 400.265625,
905
+ "completions/min_length": 287.0,
906
+ "completions/min_terminated_length": 287.0,
907
+ "entropy": 0.23906511440873146,
908
+ "epoch": 1.1772575250836121,
909
+ "frac_reward_zero_std": 0.75,
910
+ "grad_norm": 0.17765256762504578,
911
+ "kl": 0.17118846625089645,
912
+ "learning_rate": 1.2157190635451505e-06,
913
+ "loss": 0.0136,
914
+ "num_tokens": 7602903.0,
915
+ "reward": 0.875,
916
+ "reward_std": 0.12909944355487823,
917
+ "rewards/accuracy_reward/mean": 0.875,
918
+ "rewards/accuracy_reward/std": 0.2199706733226776,
919
+ "step": 1408
920
+ },
921
+ {
922
+ "clip_ratio/high_max": 8.924065696191974e-05,
923
+ "clip_ratio/high_mean": 4.462032848095987e-05,
924
+ "clip_ratio/low_mean": 4.155584974796511e-06,
925
+ "clip_ratio/low_min": 0.0,
926
+ "clip_ratio/region_mean": 4.877591345575638e-05,
927
+ "completions/clipped_ratio": 0.0,
928
+ "completions/max_length": 488.0625,
929
+ "completions/max_terminated_length": 488.0625,
930
+ "completions/mean_length": 419.373046875,
931
+ "completions/mean_terminated_length": 419.373046875,
932
+ "completions/min_length": 349.3125,
933
+ "completions/min_terminated_length": 349.3125,
934
+ "entropy": 0.2390368201304227,
935
+ "epoch": 1.2307692307692308,
936
+ "frac_reward_zero_std": 0.6875,
937
+ "grad_norm": 0.1258598417043686,
938
+ "kl": 0.15962096699513495,
939
+ "learning_rate": 1.1800445930880713e-06,
940
+ "loss": 0.0152,
941
+ "num_tokens": 7951686.0,
942
+ "reward": 0.751953125,
943
+ "reward_std": 0.11525858379900455,
944
+ "rewards/accuracy_reward/mean": 0.751953125,
945
+ "rewards/accuracy_reward/std": 0.30761036090552807,
946
+ "step": 1472
947
+ },
948
+ {
949
+ "epoch": 1.254180602006689,
950
+ "eval_completions/clipped_ratio": 0.0,
951
+ "eval_completions/max_length": 489.0,
952
+ "eval_completions/max_terminated_length": 489.0,
953
+ "eval_completions/mean_length": 402.109375,
954
+ "eval_completions/mean_terminated_length": 402.109375,
955
+ "eval_completions/min_length": 298.5,
956
+ "eval_completions/min_terminated_length": 298.5,
957
+ "eval_frac_reward_zero_std": 0.0,
958
+ "eval_loss": 0.0,
959
+ "eval_num_tokens": 8103527.0,
960
+ "eval_reward": 0.6875,
961
+ "eval_reward_std": 0.0,
962
+ "eval_rewards/accuracy_reward/mean": 0.6875,
963
+ "eval_rewards/accuracy_reward/std": 0.465938538312912,
964
+ "eval_runtime": 57.2645,
965
+ "eval_samples_per_second": 1.31,
966
+ "eval_steps_per_second": 0.035,
967
+ "step": 1500
968
+ },
969
+ {
970
+ "clip_ratio/high_max": 3.321839095507231e-05,
971
+ "clip_ratio/high_mean": 1.6609195477536155e-05,
972
+ "clip_ratio/low_mean": 0.0,
973
+ "clip_ratio/low_min": 0.0,
974
+ "clip_ratio/region_mean": 1.6609195477536155e-05,
975
+ "completions/clipped_ratio": 0.0,
976
+ "completions/max_length": 470.22222222222223,
977
+ "completions/max_terminated_length": 470.22222222222223,
978
+ "completions/mean_length": 394.00347222222223,
979
+ "completions/mean_terminated_length": 394.00347222222223,
980
+ "completions/min_length": 324.22222222222223,
981
+ "completions/min_terminated_length": 324.22222222222223,
982
+ "entropy": 0.23824424006872708,
983
+ "epoch": 1.2842809364548495,
984
+ "frac_reward_zero_std": 0.6111111111111112,
985
+ "grad_norm": 0.1226065531373024,
986
+ "kl": 0.16477220774524742,
987
+ "learning_rate": 1.1443701226309922e-06,
988
+ "loss": 0.0164,
989
+ "num_tokens": 8292504.0,
990
+ "reward": 0.6944444444444444,
991
+ "reward_std": 0.17361230651537576,
992
+ "rewards/accuracy_reward/mean": 0.6944444444444444,
993
+ "rewards/accuracy_reward/std": 0.3317638602521684,
994
+ "step": 1536
995
+ },
996
+ {
997
+ "clip_ratio/high_max": 2.729532752709929e-05,
998
+ "clip_ratio/high_mean": 1.3647663763549645e-05,
999
+ "clip_ratio/low_mean": 9.310579571319977e-06,
1000
+ "clip_ratio/low_min": 0.0,
1001
+ "clip_ratio/region_mean": 2.295824333486962e-05,
1002
+ "completions/clipped_ratio": 0.0,
1003
+ "completions/max_length": 463.8125,
1004
+ "completions/max_terminated_length": 463.8125,
1005
+ "completions/mean_length": 403.76171875,
1006
+ "completions/mean_terminated_length": 403.76171875,
1007
+ "completions/min_length": 339.0625,
1008
+ "completions/min_terminated_length": 339.0625,
1009
+ "entropy": 0.2335241783875972,
1010
+ "epoch": 1.3377926421404682,
1011
+ "frac_reward_zero_std": 0.6875,
1012
+ "grad_norm": 0.6611215472221375,
1013
+ "kl": 0.14638517287676223,
1014
+ "learning_rate": 1.108695652173913e-06,
1015
+ "loss": 0.0139,
1016
+ "num_tokens": 8633278.0,
1017
+ "reward": 0.80859375,
1018
+ "reward_std": 0.11795367486774921,
1019
+ "rewards/accuracy_reward/mean": 0.80859375,
1020
+ "rewards/accuracy_reward/std": 0.2854210138320923,
1021
+ "step": 1600
1022
+ },
1023
+ {
1024
+ "epoch": 1.3377926421404682,
1025
+ "eval_completions/clipped_ratio": 0.0,
1026
+ "eval_completions/max_length": 489.5,
1027
+ "eval_completions/max_terminated_length": 489.5,
1028
+ "eval_completions/mean_length": 411.0390625,
1029
+ "eval_completions/mean_terminated_length": 411.0390625,
1030
+ "eval_completions/min_length": 306.0,
1031
+ "eval_completions/min_terminated_length": 306.0,
1032
+ "eval_frac_reward_zero_std": 0.0,
1033
+ "eval_loss": 0.0,
1034
+ "eval_num_tokens": 8633278.0,
1035
+ "eval_reward": 0.7265625,
1036
+ "eval_reward_std": 0.0,
1037
+ "eval_rewards/accuracy_reward/mean": 0.7265625,
1038
+ "eval_rewards/accuracy_reward/std": 0.4484642744064331,
1039
+ "eval_runtime": 55.6551,
1040
+ "eval_samples_per_second": 1.348,
1041
+ "eval_steps_per_second": 0.036,
1042
+ "step": 1600
1043
+ },
1044
+ {
1045
+ "clip_ratio/high_max": 7.189887583081145e-05,
1046
+ "clip_ratio/high_mean": 4.0447135688737035e-05,
1047
+ "clip_ratio/low_mean": 4.950887159793638e-06,
1048
+ "clip_ratio/low_min": 0.0,
1049
+ "clip_ratio/region_mean": 4.539802284853067e-05,
1050
+ "completions/clipped_ratio": 0.0,
1051
+ "completions/max_length": 483.125,
1052
+ "completions/max_terminated_length": 483.125,
1053
+ "completions/mean_length": 416.267578125,
1054
+ "completions/mean_terminated_length": 416.267578125,
1055
+ "completions/min_length": 343.0625,
1056
+ "completions/min_terminated_length": 343.0625,
1057
+ "entropy": 0.2306650117971003,
1058
+ "epoch": 1.391304347826087,
1059
+ "frac_reward_zero_std": 0.78125,
1060
+ "grad_norm": 0.12204845994710922,
1061
+ "kl": 0.1382605164544657,
1062
+ "learning_rate": 1.0730211817168338e-06,
1063
+ "loss": 0.0131,
1064
+ "num_tokens": 8980599.0,
1065
+ "reward": 0.798828125,
1066
+ "reward_std": 0.08617058955132961,
1067
+ "rewards/accuracy_reward/mean": 0.798828125,
1068
+ "rewards/accuracy_reward/std": 0.25321623776108027,
1069
+ "step": 1664
1070
+ },
1071
+ {
1072
+ "epoch": 1.4214046822742474,
1073
+ "eval_completions/clipped_ratio": 0.0,
1074
+ "eval_completions/max_length": 483.0,
1075
+ "eval_completions/max_terminated_length": 483.0,
1076
+ "eval_completions/mean_length": 409.21875,
1077
+ "eval_completions/mean_terminated_length": 409.21875,
1078
+ "eval_completions/min_length": 306.5,
1079
+ "eval_completions/min_terminated_length": 306.5,
1080
+ "eval_frac_reward_zero_std": 0.0,
1081
+ "eval_loss": 0.0,
1082
+ "eval_num_tokens": 9169717.0,
1083
+ "eval_reward": 0.71875,
1084
+ "eval_reward_std": 0.0,
1085
+ "eval_rewards/accuracy_reward/mean": 0.71875,
1086
+ "eval_rewards/accuracy_reward/std": 0.4500993341207504,
1087
+ "eval_runtime": 54.3262,
1088
+ "eval_samples_per_second": 1.381,
1089
+ "eval_steps_per_second": 0.037,
1090
+ "step": 1700
1091
+ },
1092
+ {
1093
+ "clip_ratio/high_max": 2.1605737856589258e-05,
1094
+ "clip_ratio/high_mean": 1.0802868928294629e-05,
1095
+ "clip_ratio/low_mean": 0.0,
1096
+ "clip_ratio/low_min": 0.0,
1097
+ "clip_ratio/region_mean": 1.0802868928294629e-05,
1098
+ "completions/clipped_ratio": 0.0,
1099
+ "completions/max_length": 450.0,
1100
+ "completions/max_terminated_length": 450.0,
1101
+ "completions/mean_length": 392.3125,
1102
+ "completions/mean_terminated_length": 392.3125,
1103
+ "completions/min_length": 338.7142857142857,
1104
+ "completions/min_terminated_length": 338.7142857142857,
1105
+ "entropy": 0.21413783409765788,
1106
+ "epoch": 1.4448160535117056,
1107
+ "frac_reward_zero_std": 0.9285714285714286,
1108
+ "grad_norm": 0.15873290598392487,
1109
+ "kl": 0.11360293932791267,
1110
+ "learning_rate": 1.0373467112597547e-06,
1111
+ "loss": 0.0126,
1112
+ "num_tokens": 9316395.0,
1113
+ "reward": 0.9196428571428571,
1114
+ "reward_std": 0.024397500923701694,
1115
+ "rewards/accuracy_reward/mean": 0.9196428571428571,
1116
+ "rewards/accuracy_reward/std": 0.10770502473626818,
1117
+ "step": 1728
1118
+ },
1119
+ {
1120
+ "clip_ratio/high_max": 6.776595000701491e-05,
1121
+ "clip_ratio/high_mean": 3.3882975003507454e-05,
1122
+ "clip_ratio/low_mean": 4.966624146618415e-06,
1123
+ "clip_ratio/low_min": 0.0,
1124
+ "clip_ratio/region_mean": 3.884959915012587e-05,
1125
+ "completions/clipped_ratio": 0.0,
1126
+ "completions/max_length": 476.875,
1127
+ "completions/max_terminated_length": 476.875,
1128
+ "completions/mean_length": 413.81640625,
1129
+ "completions/mean_terminated_length": 413.81640625,
1130
+ "completions/min_length": 353.8125,
1131
+ "completions/min_terminated_length": 353.8125,
1132
+ "entropy": 0.2277033186983317,
1133
+ "epoch": 1.4983277591973243,
1134
+ "frac_reward_zero_std": 0.71875,
1135
+ "grad_norm": 0.18755804002285004,
1136
+ "kl": 0.15047808474628255,
1137
+ "learning_rate": 1.0016722408026757e-06,
1138
+ "loss": 0.0188,
1139
+ "num_tokens": 9662429.0,
1140
+ "reward": 0.88671875,
1141
+ "reward_std": 0.1241629384458065,
1142
+ "rewards/accuracy_reward/mean": 0.88671875,
1143
+ "rewards/accuracy_reward/std": 0.18500914610922337,
1144
+ "step": 1792
1145
+ },
1146
+ {
1147
+ "epoch": 1.5050167224080266,
1148
+ "eval_completions/clipped_ratio": 0.0,
1149
+ "eval_completions/max_length": 480.0,
1150
+ "eval_completions/max_terminated_length": 480.0,
1151
+ "eval_completions/mean_length": 403.1328125,
1152
+ "eval_completions/mean_terminated_length": 403.1328125,
1153
+ "eval_completions/min_length": 277.0,
1154
+ "eval_completions/min_terminated_length": 277.0,
1155
+ "eval_frac_reward_zero_std": 0.0,
1156
+ "eval_loss": 0.0,
1157
+ "eval_num_tokens": 9705865.0,
1158
+ "eval_reward": 0.7109375,
1159
+ "eval_reward_std": 0.0,
1160
+ "eval_rewards/accuracy_reward/mean": 0.7109375,
1161
+ "eval_rewards/accuracy_reward/std": 0.4548390805721283,
1162
+ "eval_runtime": 53.6282,
1163
+ "eval_samples_per_second": 1.399,
1164
+ "eval_steps_per_second": 0.037,
1165
+ "step": 1800
1166
+ },
1167
+ {
1168
+ "clip_ratio/high_max": 4.0922807654299375e-05,
1169
+ "clip_ratio/high_mean": 2.5359139856716085e-05,
1170
+ "clip_ratio/low_mean": 4.974134623937841e-06,
1171
+ "clip_ratio/low_min": 0.0,
1172
+ "clip_ratio/region_mean": 3.033327344123141e-05,
1173
+ "completions/clipped_ratio": 0.0,
1174
+ "completions/max_length": 471.92857142857144,
1175
+ "completions/max_terminated_length": 471.92857142857144,
1176
+ "completions/mean_length": 415.390625,
1177
+ "completions/mean_terminated_length": 415.390625,
1178
+ "completions/min_length": 353.7142857142857,
1179
+ "completions/min_terminated_length": 353.7142857142857,
1180
+ "entropy": 0.21820709614881448,
1181
+ "epoch": 1.551839464882943,
1182
+ "frac_reward_zero_std": 0.8214285714285714,
1183
+ "grad_norm": 0.9635913372039795,
1184
+ "kl": 0.10167988536081143,
1185
+ "learning_rate": 9.659977703455963e-07,
1186
+ "loss": 0.0108,
1187
+ "num_tokens": 10009128.0,
1188
+ "reward": 0.7924107142857143,
1189
+ "reward_std": 0.06061477107661111,
1190
+ "rewards/accuracy_reward/mean": 0.7924107142857143,
1191
+ "rewards/accuracy_reward/std": 0.263948740703719,
1192
+ "step": 1856
1193
+ },
1194
+ {
1195
+ "epoch": 1.588628762541806,
1196
+ "eval_completions/clipped_ratio": 0.0,
1197
+ "eval_completions/max_length": 489.0,
1198
+ "eval_completions/max_terminated_length": 489.0,
1199
+ "eval_completions/mean_length": 404.2890625,
1200
+ "eval_completions/mean_terminated_length": 404.2890625,
1201
+ "eval_completions/min_length": 275.5,
1202
+ "eval_completions/min_terminated_length": 275.5,
1203
+ "eval_frac_reward_zero_std": 0.0,
1204
+ "eval_loss": 0.0,
1205
+ "eval_num_tokens": 10243905.0,
1206
+ "eval_reward": 0.6953125,
1207
+ "eval_reward_std": 0.0,
1208
+ "eval_rewards/accuracy_reward/mean": 0.6953125,
1209
+ "eval_rewards/accuracy_reward/std": 0.46320293843746185,
1210
+ "eval_runtime": 55.0353,
1211
+ "eval_samples_per_second": 1.363,
1212
+ "eval_steps_per_second": 0.036,
1213
+ "step": 1900
1214
+ },
1215
+ {
1216
+ "clip_ratio/high_max": 0.0,
1217
+ "clip_ratio/high_mean": 0.0,
1218
+ "clip_ratio/low_mean": 0.0,
1219
+ "clip_ratio/low_min": 0.0,
1220
+ "clip_ratio/region_mean": 0.0,
1221
+ "completions/clipped_ratio": 0.0,
1222
+ "completions/max_length": 469.4,
1223
+ "completions/max_terminated_length": 469.4,
1224
+ "completions/mean_length": 421.41875,
1225
+ "completions/mean_terminated_length": 421.41875,
1226
+ "completions/min_length": 351.0,
1227
+ "completions/min_terminated_length": 351.0,
1228
+ "entropy": 0.21203163117170334,
1229
+ "epoch": 1.605351170568562,
1230
+ "frac_reward_zero_std": 1.0,
1231
+ "grad_norm": 0.15916664898395538,
1232
+ "kl": 0.1034720666706562,
1233
+ "learning_rate": 9.303232998885172e-07,
1234
+ "loss": 0.0135,
1235
+ "num_tokens": 10353220.0,
1236
+ "reward": 0.8,
1237
+ "reward_std": 0.0,
1238
+ "rewards/accuracy_reward/mean": 0.8,
1239
+ "rewards/accuracy_reward/std": 0.20320019721984864,
1240
+ "step": 1920
1241
+ },
1242
+ {
1243
+ "clip_ratio/high_max": 3.993847894889768e-05,
1244
+ "clip_ratio/high_mean": 1.996923947444884e-05,
1245
+ "clip_ratio/low_mean": 0.0,
1246
+ "clip_ratio/low_min": 0.0,
1247
+ "clip_ratio/region_mean": 1.996923947444884e-05,
1248
+ "completions/clipped_ratio": 0.0,
1249
+ "completions/max_length": 485.25,
1250
+ "completions/max_terminated_length": 485.25,
1251
+ "completions/mean_length": 421.94140625,
1252
+ "completions/mean_terminated_length": 421.94140625,
1253
+ "completions/min_length": 333.25,
1254
+ "completions/min_terminated_length": 333.25,
1255
+ "entropy": 0.22447309363633394,
1256
+ "epoch": 1.6588628762541806,
1257
+ "frac_reward_zero_std": 0.78125,
1258
+ "grad_norm": 0.1262553632259369,
1259
+ "kl": 0.1332392282783985,
1260
+ "learning_rate": 8.94648829431438e-07,
1261
+ "loss": 0.0129,
1262
+ "num_tokens": 10703478.0,
1263
+ "reward": 0.712890625,
1264
+ "reward_std": 0.09271937422454357,
1265
+ "rewards/accuracy_reward/mean": 0.712890625,
1266
+ "rewards/accuracy_reward/std": 0.28171301539987326,
1267
+ "step": 1984
1268
+ },
1269
+ {
1270
+ "epoch": 1.6722408026755853,
1271
+ "eval_completions/clipped_ratio": 0.0,
1272
+ "eval_completions/max_length": 488.0,
1273
+ "eval_completions/max_terminated_length": 488.0,
1274
+ "eval_completions/mean_length": 409.9375,
1275
+ "eval_completions/mean_terminated_length": 409.9375,
1276
+ "eval_completions/min_length": 301.5,
1277
+ "eval_completions/min_terminated_length": 301.5,
1278
+ "eval_frac_reward_zero_std": 0.0,
1279
+ "eval_loss": 0.0,
1280
+ "eval_num_tokens": 10788515.0,
1281
+ "eval_reward": 0.6953125,
1282
+ "eval_reward_std": 0.0,
1283
+ "eval_rewards/accuracy_reward/mean": 0.6953125,
1284
+ "eval_rewards/accuracy_reward/std": 0.46193525195121765,
1285
+ "eval_runtime": 55.3758,
1286
+ "eval_samples_per_second": 1.354,
1287
+ "eval_steps_per_second": 0.036,
1288
+ "step": 2000
1289
+ },
1290
+ {
1291
+ "clip_ratio/high_max": 7.684052737507348e-05,
1292
+ "clip_ratio/high_mean": 3.842026368753674e-05,
1293
+ "clip_ratio/low_mean": 0.0,
1294
+ "clip_ratio/low_min": 0.0,
1295
+ "clip_ratio/region_mean": 3.842026368753674e-05,
1296
+ "completions/clipped_ratio": 0.0,
1297
+ "completions/max_length": 466.75,
1298
+ "completions/max_terminated_length": 466.75,
1299
+ "completions/mean_length": 408.203125,
1300
+ "completions/mean_terminated_length": 408.203125,
1301
+ "completions/min_length": 340.0833333333333,
1302
+ "completions/min_terminated_length": 340.0833333333333,
1303
+ "entropy": 0.22362264028439918,
1304
+ "epoch": 1.7123745819397993,
1305
+ "frac_reward_zero_std": 0.6666666666666666,
1306
+ "grad_norm": 0.12759657204151154,
1307
+ "kl": 0.12007095486236115,
1308
+ "learning_rate": 8.589743589743588e-07,
1309
+ "loss": 0.0105,
1310
+ "num_tokens": 11045953.0,
1311
+ "reward": 0.8671875,
1312
+ "reward_std": 0.11398912469546,
1313
+ "rewards/accuracy_reward/mean": 0.8671875,
1314
+ "rewards/accuracy_reward/std": 0.22504575674732527,
1315
+ "step": 2048
1316
+ },
1317
+ {
1318
+ "epoch": 1.7558528428093645,
1319
+ "eval_completions/clipped_ratio": 0.0,
1320
+ "eval_completions/max_length": 501.5,
1321
+ "eval_completions/max_terminated_length": 501.5,
1322
+ "eval_completions/mean_length": 408.03125,
1323
+ "eval_completions/mean_terminated_length": 408.03125,
1324
+ "eval_completions/min_length": 316.0,
1325
+ "eval_completions/min_terminated_length": 316.0,
1326
+ "eval_frac_reward_zero_std": 0.0,
1327
+ "eval_loss": 0.0,
1328
+ "eval_num_tokens": 11325630.0,
1329
+ "eval_reward": 0.6796875,
1330
+ "eval_reward_std": 0.0,
1331
+ "eval_rewards/accuracy_reward/mean": 0.6796875,
1332
+ "eval_rewards/accuracy_reward/std": 0.46960316598415375,
1333
+ "eval_runtime": 56.2491,
1334
+ "eval_samples_per_second": 1.333,
1335
+ "eval_steps_per_second": 0.036,
1336
+ "step": 2100
1337
+ },
1338
+ {
1339
+ "clip_ratio/high_max": 5.1824211065346994e-05,
1340
+ "clip_ratio/high_mean": 2.5912105532673497e-05,
1341
+ "clip_ratio/low_mean": 0.0,
1342
+ "clip_ratio/low_min": 0.0,
1343
+ "clip_ratio/region_mean": 2.5912105532673497e-05,
1344
+ "completions/clipped_ratio": 0.0,
1345
+ "completions/max_length": 480.3333333333333,
1346
+ "completions/max_terminated_length": 480.3333333333333,
1347
+ "completions/mean_length": 419.3854166666667,
1348
+ "completions/mean_terminated_length": 419.3854166666667,
1349
+ "completions/min_length": 327.3333333333333,
1350
+ "completions/min_terminated_length": 327.3333333333333,
1351
+ "entropy": 0.21756159514188766,
1352
+ "epoch": 1.7658862876254182,
1353
+ "frac_reward_zero_std": 0.6666666666666666,
1354
+ "grad_norm": 0.12993162870407104,
1355
+ "kl": 0.1064981979628404,
1356
+ "learning_rate": 8.232998885172798e-07,
1357
+ "loss": 0.0143,
1358
+ "num_tokens": 11391091.0,
1359
+ "reward": 0.9791666666666666,
1360
+ "reward_std": 0.08333333333333333,
1361
+ "rewards/accuracy_reward/mean": 0.9791666666666666,
1362
+ "rewards/accuracy_reward/std": 0.11785112818082173,
1363
+ "step": 2112
1364
+ },
1365
+ {
1366
+ "clip_ratio/high_max": 9.864267667580862e-06,
1367
+ "clip_ratio/high_mean": 4.932133833790431e-06,
1368
+ "clip_ratio/low_mean": 4.852484380535316e-06,
1369
+ "clip_ratio/low_min": 0.0,
1370
+ "clip_ratio/region_mean": 9.784618214325747e-06,
1371
+ "completions/clipped_ratio": 0.0,
1372
+ "completions/max_length": 479.6875,
1373
+ "completions/max_terminated_length": 479.6875,
1374
+ "completions/mean_length": 400.6796875,
1375
+ "completions/mean_terminated_length": 400.6796875,
1376
+ "completions/min_length": 332.8125,
1377
+ "completions/min_terminated_length": 332.8125,
1378
+ "entropy": 0.21871562604792416,
1379
+ "epoch": 1.819397993311037,
1380
+ "frac_reward_zero_std": 0.75,
1381
+ "grad_norm": 0.11841385066509247,
1382
+ "kl": 0.12797342258272693,
1383
+ "learning_rate": 7.876254180602006e-07,
1384
+ "loss": 0.0127,
1385
+ "num_tokens": 11730399.0,
1386
+ "reward": 0.82421875,
1387
+ "reward_std": 0.11090045981109142,
1388
+ "rewards/accuracy_reward/mean": 0.82421875,
1389
+ "rewards/accuracy_reward/std": 0.2643054760992527,
1390
+ "step": 2176
1391
+ },
1392
+ {
1393
+ "epoch": 1.839464882943144,
1394
+ "eval_completions/clipped_ratio": 0.0,
1395
+ "eval_completions/max_length": 480.5,
1396
+ "eval_completions/max_terminated_length": 480.5,
1397
+ "eval_completions/mean_length": 409.5546875,
1398
+ "eval_completions/mean_terminated_length": 409.5546875,
1399
+ "eval_completions/min_length": 296.0,
1400
+ "eval_completions/min_terminated_length": 296.0,
1401
+ "eval_frac_reward_zero_std": 0.0,
1402
+ "eval_loss": 0.0,
1403
+ "eval_num_tokens": 11860817.0,
1404
+ "eval_reward": 0.75,
1405
+ "eval_reward_std": 0.0,
1406
+ "eval_rewards/accuracy_reward/mean": 0.75,
1407
+ "eval_rewards/accuracy_reward/std": 0.43491509556770325,
1408
+ "eval_runtime": 53.894,
1409
+ "eval_samples_per_second": 1.392,
1410
+ "eval_steps_per_second": 0.037,
1411
+ "step": 2200
1412
+ },
1413
+ {
1414
+ "clip_ratio/high_max": 4.147470317548141e-05,
1415
+ "clip_ratio/high_mean": 2.0737351587740704e-05,
1416
+ "clip_ratio/low_mean": 0.0,
1417
+ "clip_ratio/low_min": 0.0,
1418
+ "clip_ratio/region_mean": 2.0737351587740704e-05,
1419
+ "completions/clipped_ratio": 0.0,
1420
+ "completions/max_length": 486.0,
1421
+ "completions/max_terminated_length": 486.0,
1422
+ "completions/mean_length": 419.878125,
1423
+ "completions/mean_terminated_length": 419.878125,
1424
+ "completions/min_length": 339.9,
1425
+ "completions/min_terminated_length": 339.9,
1426
+ "entropy": 0.2102437574416399,
1427
+ "epoch": 1.8729096989966556,
1428
+ "frac_reward_zero_std": 0.9,
1429
+ "grad_norm": 0.09409788995981216,
1430
+ "kl": 0.09403365906327962,
1431
+ "learning_rate": 7.519509476031214e-07,
1432
+ "loss": 0.009,
1433
+ "num_tokens": 12078906.0,
1434
+ "reward": 0.79375,
1435
+ "reward_std": 0.025,
1436
+ "rewards/accuracy_reward/mean": 0.79375,
1437
+ "rewards/accuracy_reward/std": 0.23855553567409515,
1438
+ "step": 2240
1439
+ },
1440
+ {
1441
+ "epoch": 1.9230769230769231,
1442
+ "eval_completions/clipped_ratio": 0.0,
1443
+ "eval_completions/max_length": 495.0,
1444
+ "eval_completions/max_terminated_length": 495.0,
1445
+ "eval_completions/mean_length": 411.8125,
1446
+ "eval_completions/mean_terminated_length": 411.8125,
1447
+ "eval_completions/min_length": 308.0,
1448
+ "eval_completions/min_terminated_length": 308.0,
1449
+ "eval_frac_reward_zero_std": 0.0,
1450
+ "eval_loss": 0.0,
1451
+ "eval_num_tokens": 12400892.0,
1452
+ "eval_reward": 0.6875,
1453
+ "eval_reward_std": 0.0,
1454
+ "eval_rewards/accuracy_reward/mean": 0.6875,
1455
+ "eval_rewards/accuracy_reward/std": 0.46219292283058167,
1456
+ "eval_runtime": 55.4173,
1457
+ "eval_samples_per_second": 1.353,
1458
+ "eval_steps_per_second": 0.036,
1459
+ "step": 2300
1460
+ },
1461
+ {
1462
+ "clip_ratio/high_max": 0.0,
1463
+ "clip_ratio/high_mean": 0.0,
1464
+ "clip_ratio/low_mean": 0.0,
1465
+ "clip_ratio/low_min": 0.0,
1466
+ "clip_ratio/region_mean": 0.0,
1467
+ "completions/clipped_ratio": 0.0,
1468
+ "completions/max_length": 467.0,
1469
+ "completions/max_terminated_length": 467.0,
1470
+ "completions/mean_length": 437.0625,
1471
+ "completions/mean_terminated_length": 437.0625,
1472
+ "completions/min_length": 359.0,
1473
+ "completions/min_terminated_length": 359.0,
1474
+ "entropy": 0.23047670722007751,
1475
+ "epoch": 1.9264214046822743,
1476
+ "frac_reward_zero_std": 1.0,
1477
+ "grad_norm": 0.20002007484436035,
1478
+ "kl": 0.11390126869082451,
1479
+ "learning_rate": 7.162764771460424e-07,
1480
+ "loss": 0.0122,
1481
+ "num_tokens": 12423262.0,
1482
+ "reward": 1.0,
1483
+ "reward_std": 0.0,
1484
+ "rewards/accuracy_reward/mean": 1.0,
1485
+ "rewards/accuracy_reward/std": 0.0,
1486
+ "step": 2304
1487
+ },
1488
+ {
1489
+ "clip_ratio/high_max": 2.8518005819933023e-05,
1490
+ "clip_ratio/high_mean": 1.4259002909966512e-05,
1491
+ "clip_ratio/low_mean": 4.495109351410065e-06,
1492
+ "clip_ratio/low_min": 0.0,
1493
+ "clip_ratio/region_mean": 1.8754112261376576e-05,
1494
+ "completions/clipped_ratio": 0.0,
1495
+ "completions/max_length": 473.0625,
1496
+ "completions/max_terminated_length": 473.0625,
1497
+ "completions/mean_length": 413.14453125,
1498
+ "completions/mean_terminated_length": 413.14453125,
1499
+ "completions/min_length": 345.125,
1500
+ "completions/min_terminated_length": 345.125,
1501
+ "entropy": 0.21623935969546437,
1502
+ "epoch": 1.979933110367893,
1503
+ "frac_reward_zero_std": 0.8125,
1504
+ "grad_norm": 0.15597644448280334,
1505
+ "kl": 0.11163879834930412,
1506
+ "learning_rate": 6.806020066889632e-07,
1507
+ "loss": 0.0159,
1508
+ "num_tokens": 12768968.0,
1509
+ "reward": 0.677734375,
1510
+ "reward_std": 0.07654774375259876,
1511
+ "rewards/accuracy_reward/mean": 0.677734375,
1512
+ "rewards/accuracy_reward/std": 0.33232067339122295,
1513
+ "step": 2368
1514
+ },
1515
+ {
1516
+ "epoch": 2.0066889632107023,
1517
+ "eval_completions/clipped_ratio": 0.0,
1518
+ "eval_completions/max_length": 485.0,
1519
+ "eval_completions/max_terminated_length": 485.0,
1520
+ "eval_completions/mean_length": 409.28125,
1521
+ "eval_completions/mean_terminated_length": 409.28125,
1522
+ "eval_completions/min_length": 292.5,
1523
+ "eval_completions/min_terminated_length": 292.5,
1524
+ "eval_frac_reward_zero_std": 0.0,
1525
+ "eval_loss": 0.0,
1526
+ "eval_num_tokens": 12941163.0,
1527
+ "eval_reward": 0.6953125,
1528
+ "eval_reward_std": 0.0,
1529
+ "eval_rewards/accuracy_reward/mean": 0.6953125,
1530
+ "eval_rewards/accuracy_reward/std": 0.46320293843746185,
1531
+ "eval_runtime": 54.8202,
1532
+ "eval_samples_per_second": 1.368,
1533
+ "eval_steps_per_second": 0.036,
1534
+ "step": 2400
1535
+ },
1536
+ {
1537
+ "clip_ratio/high_max": 7.531194933108054e-05,
1538
+ "clip_ratio/high_mean": 3.765597466554027e-05,
1539
+ "clip_ratio/low_mean": 0.0,
1540
+ "clip_ratio/low_min": 0.0,
1541
+ "clip_ratio/region_mean": 3.765597466554027e-05,
1542
+ "completions/clipped_ratio": 0.0,
1543
+ "completions/max_length": 477.125,
1544
+ "completions/max_terminated_length": 477.125,
1545
+ "completions/mean_length": 418.265625,
1546
+ "completions/mean_terminated_length": 418.265625,
1547
+ "completions/min_length": 350.75,
1548
+ "completions/min_terminated_length": 350.75,
1549
+ "entropy": 0.21743145305663347,
1550
+ "epoch": 2.0334448160535117,
1551
+ "frac_reward_zero_std": 0.75,
1552
+ "grad_norm": 0.14243286848068237,
1553
+ "kl": 0.12919681682251394,
1554
+ "learning_rate": 6.449275362318841e-07,
1555
+ "loss": 0.011,
1556
+ "num_tokens": 13115359.0,
1557
+ "reward": 0.8125,
1558
+ "reward_std": 0.08636415377259254,
1559
+ "rewards/accuracy_reward/mean": 0.8125,
1560
+ "rewards/accuracy_reward/std": 0.2475989516824484,
1561
+ "step": 2432
1562
+ },
1563
+ {
1564
+ "clip_ratio/high_max": 4.564970004139468e-05,
1565
+ "clip_ratio/high_mean": 2.282485002069734e-05,
1566
+ "clip_ratio/low_mean": 0.0,
1567
+ "clip_ratio/low_min": 0.0,
1568
+ "clip_ratio/region_mean": 2.282485002069734e-05,
1569
+ "completions/clipped_ratio": 0.0,
1570
+ "completions/max_length": 474.5,
1571
+ "completions/max_terminated_length": 474.5,
1572
+ "completions/mean_length": 420.361328125,
1573
+ "completions/mean_terminated_length": 420.361328125,
1574
+ "completions/min_length": 346.4375,
1575
+ "completions/min_terminated_length": 346.4375,
1576
+ "entropy": 0.20763462036848068,
1577
+ "epoch": 2.0869565217391304,
1578
+ "frac_reward_zero_std": 0.875,
1579
+ "grad_norm": 0.05451633036136627,
1580
+ "kl": 0.08778653664921876,
1581
+ "learning_rate": 6.092530657748049e-07,
1582
+ "loss": 0.0081,
1583
+ "num_tokens": 13464632.0,
1584
+ "reward": 0.73828125,
1585
+ "reward_std": 0.0457378551363945,
1586
+ "rewards/accuracy_reward/mean": 0.73828125,
1587
+ "rewards/accuracy_reward/std": 0.17558534815907478,
1588
+ "step": 2496
1589
+ },
1590
+ {
1591
+ "epoch": 2.0903010033444818,
1592
+ "eval_completions/clipped_ratio": 0.0,
1593
+ "eval_completions/max_length": 487.0,
1594
+ "eval_completions/max_terminated_length": 487.0,
1595
+ "eval_completions/mean_length": 405.203125,
1596
+ "eval_completions/mean_terminated_length": 405.203125,
1597
+ "eval_completions/min_length": 277.0,
1598
+ "eval_completions/min_terminated_length": 277.0,
1599
+ "eval_frac_reward_zero_std": 0.0,
1600
+ "eval_loss": 0.0,
1601
+ "eval_num_tokens": 13486071.0,
1602
+ "eval_reward": 0.71875,
1603
+ "eval_reward_std": 0.0,
1604
+ "eval_rewards/accuracy_reward/mean": 0.71875,
1605
+ "eval_rewards/accuracy_reward/std": 0.4518062025308609,
1606
+ "eval_runtime": 55.4822,
1607
+ "eval_samples_per_second": 1.352,
1608
+ "eval_steps_per_second": 0.036,
1609
+ "step": 2500
1610
+ },
1611
+ {
1612
+ "clip_ratio/high_max": 7.478842744603753e-05,
1613
+ "clip_ratio/high_mean": 4.234863299643621e-05,
1614
+ "clip_ratio/low_mean": 0.0,
1615
+ "clip_ratio/low_min": 0.0,
1616
+ "clip_ratio/region_mean": 4.234863299643621e-05,
1617
+ "completions/clipped_ratio": 0.0,
1618
+ "completions/max_length": 472.6,
1619
+ "completions/max_terminated_length": 472.6,
1620
+ "completions/mean_length": 407.50208333333336,
1621
+ "completions/mean_terminated_length": 407.50208333333336,
1622
+ "completions/min_length": 320.0,
1623
+ "completions/min_terminated_length": 320.0,
1624
+ "entropy": 0.21952359055479367,
1625
+ "epoch": 2.140468227424749,
1626
+ "frac_reward_zero_std": 0.7333333333333333,
1627
+ "grad_norm": 0.4182104468345642,
1628
+ "kl": 0.11582428424929579,
1629
+ "learning_rate": 5.735785953177257e-07,
1630
+ "loss": 0.0094,
1631
+ "num_tokens": 13807544.0,
1632
+ "reward": 0.7583333333333333,
1633
+ "reward_std": 0.09178131222724914,
1634
+ "rewards/accuracy_reward/mean": 0.7583333333333333,
1635
+ "rewards/accuracy_reward/std": 0.2630104998747508,
1636
+ "step": 2560
1637
+ },
1638
+ {
1639
+ "epoch": 2.1739130434782608,
1640
+ "eval_completions/clipped_ratio": 0.0,
1641
+ "eval_completions/max_length": 490.0,
1642
+ "eval_completions/max_terminated_length": 490.0,
1643
+ "eval_completions/mean_length": 410.9453125,
1644
+ "eval_completions/mean_terminated_length": 410.9453125,
1645
+ "eval_completions/min_length": 314.5,
1646
+ "eval_completions/min_terminated_length": 314.5,
1647
+ "eval_frac_reward_zero_std": 0.0,
1648
+ "eval_loss": 0.0,
1649
+ "eval_num_tokens": 14028729.0,
1650
+ "eval_reward": 0.703125,
1651
+ "eval_reward_std": 0.0,
1652
+ "eval_rewards/accuracy_reward/mean": 0.703125,
1653
+ "eval_rewards/accuracy_reward/std": 0.45757468044757843,
1654
+ "eval_runtime": 55.028,
1655
+ "eval_samples_per_second": 1.363,
1656
+ "eval_steps_per_second": 0.036,
1657
+ "step": 2600
1658
+ },
1659
+ {
1660
+ "clip_ratio/high_max": 5.208333216918012e-05,
1661
+ "clip_ratio/high_mean": 2.604166608459006e-05,
1662
+ "clip_ratio/low_mean": 0.0,
1663
+ "clip_ratio/low_min": 0.0,
1664
+ "clip_ratio/region_mean": 2.604166608459006e-05,
1665
+ "completions/clipped_ratio": 0.0,
1666
+ "completions/max_length": 481.1666666666667,
1667
+ "completions/max_terminated_length": 481.1666666666667,
1668
+ "completions/mean_length": 409.8020833333333,
1669
+ "completions/mean_terminated_length": 409.8020833333333,
1670
+ "completions/min_length": 329.8333333333333,
1671
+ "completions/min_terminated_length": 329.8333333333333,
1672
+ "entropy": 0.21271449948350588,
1673
+ "epoch": 2.1939799331103678,
1674
+ "frac_reward_zero_std": 0.75,
1675
+ "grad_norm": 0.1287287324666977,
1676
+ "kl": 0.10981122233594458,
1677
+ "learning_rate": 5.379041248606467e-07,
1678
+ "loss": 0.0116,
1679
+ "num_tokens": 14157731.0,
1680
+ "reward": 0.6614583333333334,
1681
+ "reward_std": 0.10372315843900044,
1682
+ "rewards/accuracy_reward/mean": 0.6614583333333334,
1683
+ "rewards/accuracy_reward/std": 0.2969101319710414,
1684
+ "step": 2624
1685
+ },
1686
+ {
1687
+ "clip_ratio/high_max": 2.8780045795429032e-05,
1688
+ "clip_ratio/high_mean": 1.4390022897714516e-05,
1689
+ "clip_ratio/low_mean": 4.342690317571396e-06,
1690
+ "clip_ratio/low_min": 0.0,
1691
+ "clip_ratio/region_mean": 1.8732713215285912e-05,
1692
+ "completions/clipped_ratio": 0.0,
1693
+ "completions/max_length": 480.0625,
1694
+ "completions/max_terminated_length": 480.0625,
1695
+ "completions/mean_length": 421.123046875,
1696
+ "completions/mean_terminated_length": 421.123046875,
1697
+ "completions/min_length": 357.0,
1698
+ "completions/min_terminated_length": 357.0,
1699
+ "entropy": 0.21469741920009255,
1700
+ "epoch": 2.2474916387959865,
1701
+ "frac_reward_zero_std": 0.6875,
1702
+ "grad_norm": 0.14622721076011658,
1703
+ "kl": 0.12252736231312156,
1704
+ "learning_rate": 5.022296544035675e-07,
1705
+ "loss": 0.0168,
1706
+ "num_tokens": 14507554.0,
1707
+ "reward": 0.7890625,
1708
+ "reward_std": 0.11960325017571449,
1709
+ "rewards/accuracy_reward/mean": 0.7890625,
1710
+ "rewards/accuracy_reward/std": 0.25777294766157866,
1711
+ "step": 2688
1712
+ },
1713
+ {
1714
+ "epoch": 2.25752508361204,
1715
+ "eval_completions/clipped_ratio": 0.0,
1716
+ "eval_completions/max_length": 482.0,
1717
+ "eval_completions/max_terminated_length": 482.0,
1718
+ "eval_completions/mean_length": 409.46875,
1719
+ "eval_completions/mean_terminated_length": 409.46875,
1720
+ "eval_completions/min_length": 295.0,
1721
+ "eval_completions/min_terminated_length": 295.0,
1722
+ "eval_frac_reward_zero_std": 0.0,
1723
+ "eval_loss": 0.0,
1724
+ "eval_num_tokens": 14570788.0,
1725
+ "eval_reward": 0.7109375,
1726
+ "eval_reward_std": 0.0,
1727
+ "eval_rewards/accuracy_reward/mean": 0.7109375,
1728
+ "eval_rewards/accuracy_reward/std": 0.4561667740345001,
1729
+ "eval_runtime": 54.3849,
1730
+ "eval_samples_per_second": 1.379,
1731
+ "eval_steps_per_second": 0.037,
1732
+ "step": 2700
1733
+ },
1734
+ {
1735
+ "clip_ratio/high_max": 8.321751710564758e-05,
1736
+ "clip_ratio/high_mean": 4.160875855282379e-05,
1737
+ "clip_ratio/low_mean": 0.0,
1738
+ "clip_ratio/low_min": 0.0,
1739
+ "clip_ratio/region_mean": 4.160875855282379e-05,
1740
+ "completions/clipped_ratio": 0.0,
1741
+ "completions/max_length": 470.46153846153845,
1742
+ "completions/max_terminated_length": 470.46153846153845,
1743
+ "completions/mean_length": 406.8245192307692,
1744
+ "completions/mean_terminated_length": 406.8245192307692,
1745
+ "completions/min_length": 338.0769230769231,
1746
+ "completions/min_terminated_length": 338.0769230769231,
1747
+ "entropy": 0.20195185393095016,
1748
+ "epoch": 2.3010033444816056,
1749
+ "frac_reward_zero_std": 0.8076923076923077,
1750
+ "grad_norm": 0.10907543450593948,
1751
+ "kl": 0.11125540396628472,
1752
+ "learning_rate": 4.665551839464883e-07,
1753
+ "loss": 0.0098,
1754
+ "num_tokens": 14848971.0,
1755
+ "reward": 0.9206730769230769,
1756
+ "reward_std": 0.0706413984298706,
1757
+ "rewards/accuracy_reward/mean": 0.9206730769230769,
1758
+ "rewards/accuracy_reward/std": 0.14741014861143553,
1759
+ "step": 2752
1760
+ },
1761
+ {
1762
+ "epoch": 2.3411371237458196,
1763
+ "eval_completions/clipped_ratio": 0.0,
1764
+ "eval_completions/max_length": 490.5,
1765
+ "eval_completions/max_terminated_length": 490.5,
1766
+ "eval_completions/mean_length": 411.5078125,
1767
+ "eval_completions/mean_terminated_length": 411.5078125,
1768
+ "eval_completions/min_length": 291.5,
1769
+ "eval_completions/min_terminated_length": 291.5,
1770
+ "eval_frac_reward_zero_std": 0.0,
1771
+ "eval_loss": 0.0,
1772
+ "eval_num_tokens": 15102974.0,
1773
+ "eval_reward": 0.703125,
1774
+ "eval_reward_std": 0.0,
1775
+ "eval_rewards/accuracy_reward/mean": 0.703125,
1776
+ "eval_rewards/accuracy_reward/std": 0.46017006039619446,
1777
+ "eval_runtime": 55.6132,
1778
+ "eval_samples_per_second": 1.349,
1779
+ "eval_steps_per_second": 0.036,
1780
+ "step": 2800
1781
+ },
1782
+ {
1783
+ "clip_ratio/high_max": 0.0,
1784
+ "clip_ratio/high_mean": 0.0,
1785
+ "clip_ratio/low_mean": 0.0,
1786
+ "clip_ratio/low_min": 0.0,
1787
+ "clip_ratio/region_mean": 0.0,
1788
+ "completions/clipped_ratio": 0.0,
1789
+ "completions/max_length": 479.5,
1790
+ "completions/max_terminated_length": 479.5,
1791
+ "completions/mean_length": 419.90625,
1792
+ "completions/mean_terminated_length": 419.90625,
1793
+ "completions/min_length": 302.75,
1794
+ "completions/min_terminated_length": 302.75,
1795
+ "entropy": 0.20193158090114594,
1796
+ "epoch": 2.3545150501672243,
1797
+ "frac_reward_zero_std": 1.0,
1798
+ "grad_norm": 0.11240160465240479,
1799
+ "kl": 0.08153703552670777,
1800
+ "learning_rate": 4.308807134894091e-07,
1801
+ "loss": 0.0101,
1802
+ "num_tokens": 15190242.0,
1803
+ "reward": 0.625,
1804
+ "reward_std": 0.0,
1805
+ "rewards/accuracy_reward/mean": 0.625,
1806
+ "rewards/accuracy_reward/std": 0.1270001232624054,
1807
+ "step": 2816
1808
+ },
1809
+ {
1810
+ "clip_ratio/high_max": 1.9505164345900994e-05,
1811
+ "clip_ratio/high_mean": 9.752582172950497e-06,
1812
+ "clip_ratio/low_mean": 0.0,
1813
+ "clip_ratio/low_min": 0.0,
1814
+ "clip_ratio/region_mean": 9.752582172950497e-06,
1815
+ "completions/clipped_ratio": 0.0,
1816
+ "completions/max_length": 506.125,
1817
+ "completions/max_terminated_length": 506.125,
1818
+ "completions/mean_length": 423.416015625,
1819
+ "completions/mean_terminated_length": 423.416015625,
1820
+ "completions/min_length": 333.1875,
1821
+ "completions/min_terminated_length": 333.1875,
1822
+ "entropy": 0.21712527703493834,
1823
+ "epoch": 2.408026755852843,
1824
+ "frac_reward_zero_std": 0.875,
1825
+ "grad_norm": 0.1890118569135666,
1826
+ "kl": 0.10883855499560013,
1827
+ "learning_rate": 3.9520624303232996e-07,
1828
+ "loss": 0.0118,
1829
+ "num_tokens": 15541351.0,
1830
+ "reward": 0.849609375,
1831
+ "reward_std": 0.043847277760505676,
1832
+ "rewards/accuracy_reward/mean": 0.849609375,
1833
+ "rewards/accuracy_reward/std": 0.09456285089254379,
1834
+ "step": 2880
1835
+ },
1836
+ {
1837
+ "epoch": 2.4247491638795986,
1838
+ "eval_completions/clipped_ratio": 0.0,
1839
+ "eval_completions/max_length": 497.0,
1840
+ "eval_completions/max_terminated_length": 497.0,
1841
+ "eval_completions/mean_length": 410.9921875,
1842
+ "eval_completions/mean_terminated_length": 410.9921875,
1843
+ "eval_completions/min_length": 258.5,
1844
+ "eval_completions/min_terminated_length": 258.5,
1845
+ "eval_frac_reward_zero_std": 0.0,
1846
+ "eval_loss": 0.0,
1847
+ "eval_num_tokens": 15648422.0,
1848
+ "eval_reward": 0.6875,
1849
+ "eval_reward_std": 0.0,
1850
+ "eval_rewards/accuracy_reward/mean": 0.6875,
1851
+ "eval_rewards/accuracy_reward/std": 0.465938538312912,
1852
+ "eval_runtime": 56.9931,
1853
+ "eval_samples_per_second": 1.316,
1854
+ "eval_steps_per_second": 0.035,
1855
+ "step": 2900
1856
+ },
1857
+ {
1858
+ "clip_ratio/high_max": 2.8449102749370716e-05,
1859
+ "clip_ratio/high_mean": 1.4224551374685358e-05,
1860
+ "clip_ratio/low_mean": 6.500935838134452e-06,
1861
+ "clip_ratio/low_min": 0.0,
1862
+ "clip_ratio/region_mean": 2.072548721281981e-05,
1863
+ "completions/clipped_ratio": 0.0,
1864
+ "completions/max_length": 474.45454545454544,
1865
+ "completions/max_terminated_length": 474.45454545454544,
1866
+ "completions/mean_length": 420.01704545454544,
1867
+ "completions/mean_terminated_length": 420.01704545454544,
1868
+ "completions/min_length": 370.1818181818182,
1869
+ "completions/min_terminated_length": 370.1818181818182,
1870
+ "entropy": 0.20619109814817255,
1871
+ "epoch": 2.4615384615384617,
1872
+ "frac_reward_zero_std": 0.7727272727272727,
1873
+ "grad_norm": 0.18583235144615173,
1874
+ "kl": 0.1027019578455524,
1875
+ "learning_rate": 3.595317725752508e-07,
1876
+ "loss": 0.0091,
1877
+ "num_tokens": 15888348.0,
1878
+ "reward": 0.8892045454545454,
1879
+ "reward_std": 0.08273993567986922,
1880
+ "rewards/accuracy_reward/mean": 0.8892045454545454,
1881
+ "rewards/accuracy_reward/std": 0.15522026880220932,
1882
+ "step": 2944
1883
+ },
1884
+ {
1885
+ "epoch": 2.508361204013378,
1886
+ "eval_completions/clipped_ratio": 0.0,
1887
+ "eval_completions/max_length": 484.5,
1888
+ "eval_completions/max_terminated_length": 484.5,
1889
+ "eval_completions/mean_length": 407.7890625,
1890
+ "eval_completions/mean_terminated_length": 407.7890625,
1891
+ "eval_completions/min_length": 260.0,
1892
+ "eval_completions/min_terminated_length": 260.0,
1893
+ "eval_frac_reward_zero_std": 0.0,
1894
+ "eval_loss": 0.0,
1895
+ "eval_num_tokens": 16190533.0,
1896
+ "eval_reward": 0.703125,
1897
+ "eval_reward_std": 0.0,
1898
+ "eval_rewards/accuracy_reward/mean": 0.703125,
1899
+ "eval_rewards/accuracy_reward/std": 0.4591996520757675,
1900
+ "eval_runtime": 55.9308,
1901
+ "eval_samples_per_second": 1.341,
1902
+ "eval_steps_per_second": 0.036,
1903
+ "step": 3000
1904
+ },
1905
+ {
1906
+ "clip_ratio/high_max": 0.0,
1907
+ "clip_ratio/high_mean": 0.0,
1908
+ "clip_ratio/low_mean": 0.0,
1909
+ "clip_ratio/low_min": 0.0,
1910
+ "clip_ratio/region_mean": 0.0,
1911
+ "completions/clipped_ratio": 0.0,
1912
+ "completions/max_length": 485.5,
1913
+ "completions/max_terminated_length": 485.5,
1914
+ "completions/mean_length": 397.703125,
1915
+ "completions/mean_terminated_length": 397.703125,
1916
+ "completions/min_length": 323.0,
1917
+ "completions/min_terminated_length": 323.0,
1918
+ "entropy": 0.1908447314053774,
1919
+ "epoch": 2.5150501672240804,
1920
+ "frac_reward_zero_std": 0.5,
1921
+ "grad_norm": 0.294365257024765,
1922
+ "kl": 0.06209766957908869,
1923
+ "learning_rate": 3.238573021181717e-07,
1924
+ "loss": 0.0104,
1925
+ "num_tokens": 16232754.0,
1926
+ "reward": 0.96875,
1927
+ "reward_std": 0.125,
1928
+ "rewards/accuracy_reward/mean": 0.96875,
1929
+ "rewards/accuracy_reward/std": 0.1767766922712326,
1930
+ "step": 3008
1931
+ },
1932
+ {
1933
+ "clip_ratio/high_max": 2.0032051907037385e-05,
1934
+ "clip_ratio/high_mean": 1.0016025953518692e-05,
1935
+ "clip_ratio/low_mean": 0.0,
1936
+ "clip_ratio/low_min": 0.0,
1937
+ "clip_ratio/region_mean": 1.0016025953518692e-05,
1938
+ "completions/clipped_ratio": 0.0,
1939
+ "completions/max_length": 474.75,
1940
+ "completions/max_terminated_length": 474.75,
1941
+ "completions/mean_length": 414.232421875,
1942
+ "completions/mean_terminated_length": 414.232421875,
1943
+ "completions/min_length": 342.9375,
1944
+ "completions/min_terminated_length": 342.9375,
1945
+ "entropy": 0.20259337592869997,
1946
+ "epoch": 2.568561872909699,
1947
+ "frac_reward_zero_std": 0.8125,
1948
+ "grad_norm": 0.11983966082334518,
1949
+ "kl": 0.08080932262237184,
1950
+ "learning_rate": 2.881828316610925e-07,
1951
+ "loss": 0.0084,
1952
+ "num_tokens": 16578905.0,
1953
+ "reward": 0.787109375,
1954
+ "reward_std": 0.05452118441462517,
1955
+ "rewards/accuracy_reward/mean": 0.787109375,
1956
+ "rewards/accuracy_reward/std": 0.1910215886309743,
1957
+ "step": 3072
1958
+ },
1959
+ {
1960
+ "epoch": 2.591973244147157,
1961
+ "eval_completions/clipped_ratio": 0.0,
1962
+ "eval_completions/max_length": 494.0,
1963
+ "eval_completions/max_terminated_length": 494.0,
1964
+ "eval_completions/mean_length": 413.71875,
1965
+ "eval_completions/mean_terminated_length": 413.71875,
1966
+ "eval_completions/min_length": 321.5,
1967
+ "eval_completions/min_terminated_length": 321.5,
1968
+ "eval_frac_reward_zero_std": 0.0,
1969
+ "eval_loss": 0.0,
1970
+ "eval_num_tokens": 16730197.0,
1971
+ "eval_reward": 0.703125,
1972
+ "eval_reward_std": 0.0,
1973
+ "eval_rewards/accuracy_reward/mean": 0.703125,
1974
+ "eval_rewards/accuracy_reward/std": 0.45757468044757843,
1975
+ "eval_runtime": 56.085,
1976
+ "eval_samples_per_second": 1.337,
1977
+ "eval_steps_per_second": 0.036,
1978
+ "step": 3100
1979
+ },
1980
+ {
1981
+ "clip_ratio/high_max": 0.0,
1982
+ "clip_ratio/high_mean": 0.0,
1983
+ "clip_ratio/low_mean": 0.0,
1984
+ "clip_ratio/low_min": 0.0,
1985
+ "clip_ratio/region_mean": 0.0,
1986
+ "completions/clipped_ratio": 0.0,
1987
+ "completions/max_length": 477.6666666666667,
1988
+ "completions/max_terminated_length": 477.6666666666667,
1989
+ "completions/mean_length": 407.5729166666667,
1990
+ "completions/mean_terminated_length": 407.5729166666667,
1991
+ "completions/min_length": 336.8888888888889,
1992
+ "completions/min_terminated_length": 336.8888888888889,
1993
+ "entropy": 0.2053673474325074,
1994
+ "epoch": 2.6220735785953178,
1995
+ "frac_reward_zero_std": 0.8333333333333334,
1996
+ "grad_norm": 0.09832081943750381,
1997
+ "kl": 0.08688689861446619,
1998
+ "learning_rate": 2.5250836120401335e-07,
1999
+ "loss": 0.0105,
2000
+ "num_tokens": 16923050.0,
2001
+ "reward": 0.7395833333333334,
2002
+ "reward_std": 0.05017293824089898,
2003
+ "rewards/accuracy_reward/mean": 0.7395833333333334,
2004
+ "rewards/accuracy_reward/std": 0.20655767785178292,
2005
+ "step": 3136
2006
+ },
2007
+ {
2008
+ "clip_ratio/high_max": 8.985049134935252e-06,
2009
+ "clip_ratio/high_mean": 4.492524567467626e-06,
2010
+ "clip_ratio/low_mean": 0.0,
2011
+ "clip_ratio/low_min": 0.0,
2012
+ "clip_ratio/region_mean": 4.492524567467626e-06,
2013
+ "completions/clipped_ratio": 0.0,
2014
+ "completions/max_length": 479.0,
2015
+ "completions/max_terminated_length": 479.0,
2016
+ "completions/mean_length": 424.021484375,
2017
+ "completions/mean_terminated_length": 424.021484375,
2018
+ "completions/min_length": 347.125,
2019
+ "completions/min_terminated_length": 347.125,
2020
+ "entropy": 0.20460248668678105,
2021
+ "epoch": 2.6755852842809364,
2022
+ "frac_reward_zero_std": 0.875,
2023
+ "grad_norm": 0.14690326154232025,
2024
+ "kl": 0.09142500828602351,
2025
+ "learning_rate": 2.1683389074693424e-07,
2026
+ "loss": 0.0088,
2027
+ "num_tokens": 17274293.0,
2028
+ "reward": 0.765625,
2029
+ "reward_std": 0.04243633709847927,
2030
+ "rewards/accuracy_reward/mean": 0.765625,
2031
+ "rewards/accuracy_reward/std": 0.18196979723870754,
2032
+ "step": 3200
2033
+ },
2034
+ {
2035
+ "epoch": 2.6755852842809364,
2036
+ "eval_completions/clipped_ratio": 0.0,
2037
+ "eval_completions/max_length": 475.5,
2038
+ "eval_completions/max_terminated_length": 475.5,
2039
+ "eval_completions/mean_length": 407.9765625,
2040
+ "eval_completions/mean_terminated_length": 407.9765625,
2041
+ "eval_completions/min_length": 306.5,
2042
+ "eval_completions/min_terminated_length": 306.5,
2043
+ "eval_frac_reward_zero_std": 0.0,
2044
+ "eval_loss": 0.0,
2045
+ "eval_num_tokens": 17274293.0,
2046
+ "eval_reward": 0.6953125,
2047
+ "eval_reward_std": 0.0,
2048
+ "eval_rewards/accuracy_reward/mean": 0.6953125,
2049
+ "eval_rewards/accuracy_reward/std": 0.4638346880674362,
2050
+ "eval_runtime": 53.8229,
2051
+ "eval_samples_per_second": 1.393,
2052
+ "eval_steps_per_second": 0.037,
2053
+ "step": 3200
2054
+ },
2055
+ {
2056
+ "clip_ratio/high_max": 9.895503353618551e-06,
2057
+ "clip_ratio/high_mean": 4.9477516768092755e-06,
2058
+ "clip_ratio/low_mean": 4.5767428673570976e-06,
2059
+ "clip_ratio/low_min": 0.0,
2060
+ "clip_ratio/region_mean": 9.524494544166373e-06,
2061
+ "completions/clipped_ratio": 0.0,
2062
+ "completions/max_length": 471.1875,
2063
+ "completions/max_terminated_length": 471.1875,
2064
+ "completions/mean_length": 406.287109375,
2065
+ "completions/mean_terminated_length": 406.287109375,
2066
+ "completions/min_length": 330.75,
2067
+ "completions/min_terminated_length": 330.75,
2068
+ "entropy": 0.193967598490417,
2069
+ "epoch": 2.729096989966555,
2070
+ "frac_reward_zero_std": 0.90625,
2071
+ "grad_norm": 0.15071481466293335,
2072
+ "kl": 0.08262389054289088,
2073
+ "learning_rate": 1.8115942028985507e-07,
2074
+ "loss": 0.007,
2075
+ "num_tokens": 17616488.0,
2076
+ "reward": 0.90625,
2077
+ "reward_std": 0.03996079042553902,
2078
+ "rewards/accuracy_reward/mean": 0.90625,
2079
+ "rewards/accuracy_reward/std": 0.11948453821241856,
2080
+ "step": 3264
2081
+ },
2082
+ {
2083
+ "epoch": 2.759197324414716,
2084
+ "eval_completions/clipped_ratio": 0.0,
2085
+ "eval_completions/max_length": 482.0,
2086
+ "eval_completions/max_terminated_length": 482.0,
2087
+ "eval_completions/mean_length": 409.1171875,
2088
+ "eval_completions/mean_terminated_length": 409.1171875,
2089
+ "eval_completions/min_length": 288.5,
2090
+ "eval_completions/min_terminated_length": 288.5,
2091
+ "eval_frac_reward_zero_std": 0.0,
2092
+ "eval_loss": 0.0,
2093
+ "eval_num_tokens": 17808921.0,
2094
+ "eval_reward": 0.7109375,
2095
+ "eval_reward_std": 0.0,
2096
+ "eval_rewards/accuracy_reward/mean": 0.7109375,
2097
+ "eval_rewards/accuracy_reward/std": 0.4561667740345001,
2098
+ "eval_runtime": 54.9248,
2099
+ "eval_samples_per_second": 1.366,
2100
+ "eval_steps_per_second": 0.036,
2101
+ "step": 3300
2102
+ },
2103
+ {
2104
+ "clip_ratio/high_max": 4.422420793811658e-05,
2105
+ "clip_ratio/high_mean": 2.211210396905829e-05,
2106
+ "clip_ratio/low_mean": 1.0725010984710284e-05,
2107
+ "clip_ratio/low_min": 0.0,
2108
+ "clip_ratio/region_mean": 3.2837114953768575e-05,
2109
+ "completions/clipped_ratio": 0.0,
2110
+ "completions/max_length": 471.7142857142857,
2111
+ "completions/max_terminated_length": 471.7142857142857,
2112
+ "completions/mean_length": 405.16964285714283,
2113
+ "completions/mean_terminated_length": 405.16964285714283,
2114
+ "completions/min_length": 344.85714285714283,
2115
+ "completions/min_terminated_length": 344.85714285714283,
2116
+ "entropy": 0.20200447046330997,
2117
+ "epoch": 2.782608695652174,
2118
+ "frac_reward_zero_std": 0.6428571428571429,
2119
+ "grad_norm": 0.13082604110240936,
2120
+ "kl": 0.1005759423465601,
2121
+ "learning_rate": 1.454849498327759e-07,
2122
+ "loss": 0.0146,
2123
+ "num_tokens": 17958399.0,
2124
+ "reward": 0.7142857142857143,
2125
+ "reward_std": 0.1647304126194545,
2126
+ "rewards/accuracy_reward/mean": 0.7142857142857143,
2127
+ "rewards/accuracy_reward/std": 0.260468738419669,
2128
+ "step": 3328
2129
+ },
2130
+ {
2131
+ "clip_ratio/high_max": 1.0028883480117656e-05,
2132
+ "clip_ratio/high_mean": 5.014441740058828e-06,
2133
+ "clip_ratio/low_mean": 4.266793894203147e-06,
2134
+ "clip_ratio/low_min": 0.0,
2135
+ "clip_ratio/region_mean": 9.281235634261975e-06,
2136
+ "completions/clipped_ratio": 0.0,
2137
+ "completions/max_length": 481.8125,
2138
+ "completions/max_terminated_length": 481.8125,
2139
+ "completions/mean_length": 418.5859375,
2140
+ "completions/mean_terminated_length": 418.5859375,
2141
+ "completions/min_length": 341.9375,
2142
+ "completions/min_terminated_length": 341.9375,
2143
+ "entropy": 0.20107114291749895,
2144
+ "epoch": 2.8361204013377925,
2145
+ "frac_reward_zero_std": 0.90625,
2146
+ "grad_norm": 0.18976449966430664,
2147
+ "kl": 0.08409578536520712,
2148
+ "learning_rate": 1.0981047937569676e-07,
2149
+ "loss": 0.0092,
2150
+ "num_tokens": 18306923.0,
2151
+ "reward": 0.8671875,
2152
+ "reward_std": 0.02960042469203472,
2153
+ "rewards/accuracy_reward/mean": 0.8671875,
2154
+ "rewards/accuracy_reward/std": 0.12586778588593006,
2155
+ "step": 3392
2156
+ },
2157
+ {
2158
+ "epoch": 2.842809364548495,
2159
+ "eval_completions/clipped_ratio": 0.0,
2160
+ "eval_completions/max_length": 480.5,
2161
+ "eval_completions/max_terminated_length": 480.5,
2162
+ "eval_completions/mean_length": 409.984375,
2163
+ "eval_completions/mean_terminated_length": 409.984375,
2164
+ "eval_completions/min_length": 312.0,
2165
+ "eval_completions/min_terminated_length": 312.0,
2166
+ "eval_frac_reward_zero_std": 0.0,
2167
+ "eval_loss": 0.0,
2168
+ "eval_num_tokens": 18348241.0,
2169
+ "eval_reward": 0.6875,
2170
+ "eval_reward_std": 0.0,
2171
+ "eval_rewards/accuracy_reward/mean": 0.6875,
2172
+ "eval_rewards/accuracy_reward/std": 0.46219292283058167,
2173
+ "eval_runtime": 54.1878,
2174
+ "eval_samples_per_second": 1.384,
2175
+ "eval_steps_per_second": 0.037,
2176
+ "step": 3400
2177
+ },
2178
+ {
2179
+ "clip_ratio/high_max": 2.266245220588254e-05,
2180
+ "clip_ratio/high_mean": 1.133122610294127e-05,
2181
+ "clip_ratio/low_mean": 5.305152236750084e-06,
2182
+ "clip_ratio/low_min": 0.0,
2183
+ "clip_ratio/region_mean": 1.6636378339691354e-05,
2184
+ "completions/clipped_ratio": 0.0,
2185
+ "completions/max_length": 482.2142857142857,
2186
+ "completions/max_terminated_length": 482.2142857142857,
2187
+ "completions/mean_length": 414.65401785714283,
2188
+ "completions/mean_terminated_length": 414.65401785714283,
2189
+ "completions/min_length": 347.7142857142857,
2190
+ "completions/min_terminated_length": 347.7142857142857,
2191
+ "entropy": 0.20981859575424874,
2192
+ "epoch": 2.8896321070234112,
2193
+ "frac_reward_zero_std": 0.7857142857142857,
2194
+ "grad_norm": 1.1070315837860107,
2195
+ "kl": 0.10563170131562012,
2196
+ "learning_rate": 7.41360089186176e-08,
2197
+ "loss": 0.0135,
2198
+ "num_tokens": 18651446.0,
2199
+ "reward": 0.7857142857142857,
2200
+ "reward_std": 0.0999031024319785,
2201
+ "rewards/accuracy_reward/mean": 0.7857142857142857,
2202
+ "rewards/accuracy_reward/std": 0.2894010437386377,
2203
+ "step": 3456
2204
+ },
2205
+ {
2206
+ "epoch": 2.9264214046822743,
2207
+ "eval_completions/clipped_ratio": 0.0,
2208
+ "eval_completions/max_length": 479.5,
2209
+ "eval_completions/max_terminated_length": 479.5,
2210
+ "eval_completions/mean_length": 409.65625,
2211
+ "eval_completions/mean_terminated_length": 409.65625,
2212
+ "eval_completions/min_length": 302.5,
2213
+ "eval_completions/min_terminated_length": 302.5,
2214
+ "eval_frac_reward_zero_std": 0.0,
2215
+ "eval_loss": 0.0,
2216
+ "eval_num_tokens": 18887628.0,
2217
+ "eval_reward": 0.6953125,
2218
+ "eval_reward_std": 0.0,
2219
+ "eval_rewards/accuracy_reward/mean": 0.6953125,
2220
+ "eval_rewards/accuracy_reward/std": 0.46193525195121765,
2221
+ "eval_runtime": 53.6776,
2222
+ "eval_samples_per_second": 1.397,
2223
+ "eval_steps_per_second": 0.037,
2224
+ "step": 3500
2225
+ },
2226
+ {
2227
+ "clip_ratio/high_max": 6.487761856988072e-05,
2228
+ "clip_ratio/high_mean": 3.243880928494036e-05,
2229
+ "clip_ratio/low_mean": 0.0,
2230
+ "clip_ratio/low_min": 0.0,
2231
+ "clip_ratio/region_mean": 3.243880928494036e-05,
2232
+ "completions/clipped_ratio": 0.0,
2233
+ "completions/max_length": 465.0,
2234
+ "completions/max_terminated_length": 465.0,
2235
+ "completions/mean_length": 402.0875,
2236
+ "completions/mean_terminated_length": 402.0875,
2237
+ "completions/min_length": 321.8,
2238
+ "completions/min_terminated_length": 321.8,
2239
+ "entropy": 0.2007051318883896,
2240
+ "epoch": 2.94314381270903,
2241
+ "frac_reward_zero_std": 0.9,
2242
+ "grad_norm": 0.10725326836109161,
2243
+ "kl": 0.09607526948675513,
2244
+ "learning_rate": 3.846153846153846e-08,
2245
+ "loss": 0.0081,
2246
+ "num_tokens": 18993898.0,
2247
+ "reward": 0.76875,
2248
+ "reward_std": 0.047871357202529906,
2249
+ "rewards/accuracy_reward/mean": 0.76875,
2250
+ "rewards/accuracy_reward/std": 0.1981118381023407,
2251
+ "step": 3520
2252
+ },
2253
+ {
2254
+ "clip_ratio/high_max": 2.1599041247100104e-05,
2255
+ "clip_ratio/high_mean": 1.0799520623550052e-05,
2256
+ "clip_ratio/low_mean": 0.0,
2257
+ "clip_ratio/low_min": 0.0,
2258
+ "clip_ratio/region_mean": 1.0799520623550052e-05,
2259
+ "completions/clipped_ratio": 0.0,
2260
+ "completions/max_length": 480.0,
2261
+ "completions/max_terminated_length": 480.0,
2262
+ "completions/mean_length": 411.005859375,
2263
+ "completions/mean_terminated_length": 411.005859375,
2264
+ "completions/min_length": 343.625,
2265
+ "completions/min_terminated_length": 343.625,
2266
+ "entropy": 0.2045821927022189,
2267
+ "epoch": 2.9966555183946486,
2268
+ "frac_reward_zero_std": 0.84375,
2269
+ "grad_norm": 0.10782357305288315,
2270
+ "kl": 0.08409121469594538,
2271
+ "learning_rate": 2.787068004459309e-09,
2272
+ "loss": 0.0098,
2273
+ "num_tokens": 19338349.0,
2274
+ "reward": 0.880859375,
2275
+ "reward_std": 0.05523413047194481,
2276
+ "rewards/accuracy_reward/mean": 0.880859375,
2277
+ "rewards/accuracy_reward/std": 0.17677514534443617,
2278
+ "step": 3584
2279
+ },
2280
+ {
2281
+ "clip_ratio/high_max": 0.0,
2282
+ "clip_ratio/high_mean": 0.0,
2283
+ "clip_ratio/low_mean": 0.0,
2284
+ "clip_ratio/low_min": 0.0,
2285
+ "clip_ratio/region_mean": 0.0,
2286
+ "completions/clipped_ratio": 0.0,
2287
+ "completions/max_length": 484.0,
2288
+ "completions/max_terminated_length": 484.0,
2289
+ "completions/mean_length": 422.125,
2290
+ "completions/mean_terminated_length": 422.125,
2291
+ "completions/min_length": 360.0,
2292
+ "completions/min_terminated_length": 360.0,
2293
+ "entropy": 0.20116707310080528,
2294
+ "epoch": 3.0,
2295
+ "frac_reward_zero_std": 1.0,
2296
+ "kl": 0.06294638197869062,
2297
+ "num_tokens": 19360225.0,
2298
+ "reward": 0.5,
2299
+ "reward_std": 0.0,
2300
+ "rewards/accuracy_reward/mean": 0.5,
2301
+ "rewards/accuracy_reward/std": 0.5080004930496216,
2302
+ "step": 3588,
2303
+ "total_flos": 0.0,
2304
+ "train_loss": 0.010405275567721289,
2305
+ "train_runtime": 34829.3301,
2306
+ "train_samples_per_second": 0.052,
2307
+ "train_steps_per_second": 0.103
2308
+ }
2309
+ ],
2310
+ "logging_steps": 64,
2311
+ "max_steps": 3588,
2312
+ "num_input_tokens_seen": 19360225,
2313
+ "num_train_epochs": 3,
2314
+ "save_steps": 100,
2315
+ "stateful_callbacks": {
2316
+ "TrainerControl": {
2317
+ "args": {
2318
+ "should_epoch_stop": false,
2319
+ "should_evaluate": false,
2320
+ "should_log": false,
2321
+ "should_save": true,
2322
+ "should_training_stop": true
2323
+ },
2324
+ "attributes": {}
2325
+ }
2326
+ },
2327
+ "total_flos": 0.0,
2328
+ "train_batch_size": 4,
2329
+ "trial_name": null,
2330
+ "trial_params": null
2331
+ }
Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 11008,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention"
49
+ ],
50
+ "max_position_embeddings": 32768,
51
+ "max_window_layers": 70,
52
+ "model_type": "qwen2",
53
+ "num_attention_heads": 16,
54
+ "num_hidden_layers": 36,
55
+ "num_key_value_heads": 2,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.55.2",
63
+ "use_cache": true,
64
+ "use_sliding_window": false,
65
+ "vocab_size": 151936
66
+ }
Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.55.2"
14
+ }
Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/model.safetensors.index.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_parameters": 3085938688,
4
+ "total_size": 6794207232
5
+ },
6
+ "weight_map": {
7
+ "lm_head.weight": "model-00002-of-00002.safetensors",
8
+ "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
9
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
10
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
11
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
12
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
13
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
14
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
15
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
16
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
17
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
18
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
19
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
20
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
21
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors",
22
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
23
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
24
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
25
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
26
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
27
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
28
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
29
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
30
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
31
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
32
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
33
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors",
34
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
35
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
36
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
37
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
38
+ "model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
39
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
40
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
41
+ "model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
42
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
43
+ "model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
44
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
45
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors",
46
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
47
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
48
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
49
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
50
+ "model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
51
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
52
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
53
+ "model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
54
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
55
+ "model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
56
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
57
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors",
58
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
59
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
60
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
61
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
62
+ "model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
63
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
64
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
65
+ "model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
66
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
67
+ "model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
68
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
69
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
70
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
71
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
72
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
73
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
74
+ "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
75
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
76
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
77
+ "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
78
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
79
+ "model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
80
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
81
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
82
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
83
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
84
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
85
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
86
+ "model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
87
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
88
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
89
+ "model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
90
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
91
+ "model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
92
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
93
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors",
94
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
95
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
96
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
97
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
98
+ "model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
99
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
100
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
101
+ "model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
102
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
103
+ "model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
104
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
105
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors",
106
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
107
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
108
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
109
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
110
+ "model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
111
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
112
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
113
+ "model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
114
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
115
+ "model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
116
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
117
+ "model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors",
118
+ "model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
119
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
120
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
121
+ "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
122
+ "model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
123
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
124
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
125
+ "model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
126
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
127
+ "model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
128
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
129
+ "model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
130
+ "model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
131
+ "model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
132
+ "model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
133
+ "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
134
+ "model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
135
+ "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
136
+ "model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
137
+ "model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
138
+ "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
139
+ "model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
140
+ "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
141
+ "model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
142
+ "model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
143
+ "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
144
+ "model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
145
+ "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
146
+ "model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
147
+ "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
148
+ "model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
149
+ "model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
150
+ "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
151
+ "model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
152
+ "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
153
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
154
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
155
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
156
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
157
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
158
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
159
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
160
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
161
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
162
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
163
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
164
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
165
+ "model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
166
+ "model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
167
+ "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
168
+ "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
169
+ "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
170
+ "model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
171
+ "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
172
+ "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
173
+ "model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
174
+ "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
175
+ "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
176
+ "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
177
+ "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
178
+ "model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
179
+ "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
180
+ "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
181
+ "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
182
+ "model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
183
+ "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
184
+ "model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
185
+ "model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
186
+ "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
187
+ "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
188
+ "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
189
+ "model.layers.22.input_layernorm.weight": "model-00001-of-00002.safetensors",
190
+ "model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
191
+ "model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
192
+ "model.layers.22.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
193
+ "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
194
+ "model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
195
+ "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
196
+ "model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
197
+ "model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
198
+ "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
199
+ "model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
200
+ "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
201
+ "model.layers.23.input_layernorm.weight": "model-00001-of-00002.safetensors",
202
+ "model.layers.23.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
203
+ "model.layers.23.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
204
+ "model.layers.23.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
205
+ "model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
206
+ "model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
207
+ "model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
208
+ "model.layers.23.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
209
+ "model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
210
+ "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
211
+ "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
212
+ "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
213
+ "model.layers.24.input_layernorm.weight": "model-00001-of-00002.safetensors",
214
+ "model.layers.24.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
215
+ "model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
216
+ "model.layers.24.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
217
+ "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
218
+ "model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
219
+ "model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
220
+ "model.layers.24.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
221
+ "model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
222
+ "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
223
+ "model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
224
+ "model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
225
+ "model.layers.25.input_layernorm.weight": "model-00001-of-00002.safetensors",
226
+ "model.layers.25.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
227
+ "model.layers.25.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
228
+ "model.layers.25.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
229
+ "model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
230
+ "model.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
231
+ "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
232
+ "model.layers.25.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
233
+ "model.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
234
+ "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
235
+ "model.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
236
+ "model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
237
+ "model.layers.26.input_layernorm.weight": "model-00001-of-00002.safetensors",
238
+ "model.layers.26.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
239
+ "model.layers.26.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
240
+ "model.layers.26.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
241
+ "model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
242
+ "model.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
243
+ "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
244
+ "model.layers.26.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
245
+ "model.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
246
+ "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
247
+ "model.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
248
+ "model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
249
+ "model.layers.27.input_layernorm.weight": "model-00001-of-00002.safetensors",
250
+ "model.layers.27.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
251
+ "model.layers.27.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
252
+ "model.layers.27.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
253
+ "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
254
+ "model.layers.27.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
255
+ "model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
256
+ "model.layers.27.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
257
+ "model.layers.27.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
258
+ "model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
259
+ "model.layers.27.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
260
+ "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
261
+ "model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors",
262
+ "model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
263
+ "model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
264
+ "model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
265
+ "model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
266
+ "model.layers.28.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
267
+ "model.layers.28.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
268
+ "model.layers.28.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
269
+ "model.layers.28.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
270
+ "model.layers.28.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
271
+ "model.layers.28.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
272
+ "model.layers.28.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
273
+ "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
274
+ "model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
275
+ "model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
276
+ "model.layers.29.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
277
+ "model.layers.29.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
278
+ "model.layers.29.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
279
+ "model.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
280
+ "model.layers.29.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
281
+ "model.layers.29.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
282
+ "model.layers.29.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
283
+ "model.layers.29.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
284
+ "model.layers.29.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
285
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
286
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
287
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
288
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
289
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
290
+ "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
291
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
292
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
293
+ "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
294
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
295
+ "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
296
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
297
+ "model.layers.30.input_layernorm.weight": "model-00002-of-00002.safetensors",
298
+ "model.layers.30.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
299
+ "model.layers.30.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
300
+ "model.layers.30.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
301
+ "model.layers.30.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
302
+ "model.layers.30.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
303
+ "model.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
304
+ "model.layers.30.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
305
+ "model.layers.30.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
306
+ "model.layers.30.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
307
+ "model.layers.30.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
308
+ "model.layers.30.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
309
+ "model.layers.31.input_layernorm.weight": "model-00002-of-00002.safetensors",
310
+ "model.layers.31.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
311
+ "model.layers.31.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
312
+ "model.layers.31.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
313
+ "model.layers.31.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
314
+ "model.layers.31.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
315
+ "model.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
316
+ "model.layers.31.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
317
+ "model.layers.31.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
318
+ "model.layers.31.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
319
+ "model.layers.31.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
320
+ "model.layers.31.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
321
+ "model.layers.32.input_layernorm.weight": "model-00002-of-00002.safetensors",
322
+ "model.layers.32.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
323
+ "model.layers.32.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
324
+ "model.layers.32.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
325
+ "model.layers.32.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
326
+ "model.layers.32.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
327
+ "model.layers.32.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
328
+ "model.layers.32.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
329
+ "model.layers.32.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
330
+ "model.layers.32.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
331
+ "model.layers.32.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
332
+ "model.layers.32.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
333
+ "model.layers.33.input_layernorm.weight": "model-00002-of-00002.safetensors",
334
+ "model.layers.33.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
335
+ "model.layers.33.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
336
+ "model.layers.33.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
337
+ "model.layers.33.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
338
+ "model.layers.33.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
339
+ "model.layers.33.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
340
+ "model.layers.33.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
341
+ "model.layers.33.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
342
+ "model.layers.33.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
343
+ "model.layers.33.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
344
+ "model.layers.33.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
345
+ "model.layers.34.input_layernorm.weight": "model-00002-of-00002.safetensors",
346
+ "model.layers.34.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
347
+ "model.layers.34.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
348
+ "model.layers.34.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
349
+ "model.layers.34.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
350
+ "model.layers.34.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
351
+ "model.layers.34.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
352
+ "model.layers.34.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
353
+ "model.layers.34.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
354
+ "model.layers.34.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
355
+ "model.layers.34.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
356
+ "model.layers.34.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
357
+ "model.layers.35.input_layernorm.weight": "model-00002-of-00002.safetensors",
358
+ "model.layers.35.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
359
+ "model.layers.35.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
360
+ "model.layers.35.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
361
+ "model.layers.35.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
362
+ "model.layers.35.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
363
+ "model.layers.35.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
364
+ "model.layers.35.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
365
+ "model.layers.35.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
366
+ "model.layers.35.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
367
+ "model.layers.35.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
368
+ "model.layers.35.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
369
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
370
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
371
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
372
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
373
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
374
+ "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
375
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
376
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
377
+ "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
378
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
379
+ "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
380
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
381
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
382
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
383
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
384
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
385
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
386
+ "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
387
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
388
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
389
+ "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
390
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
391
+ "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
392
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
393
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
394
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
395
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
396
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
397
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
398
+ "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
399
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
400
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
401
+ "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
402
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
403
+ "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
404
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
405
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors",
406
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
407
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
408
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
409
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
410
+ "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
411
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
412
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
413
+ "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
414
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
415
+ "model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
416
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
417
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors",
418
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
419
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
420
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
421
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
422
+ "model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
423
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
424
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
425
+ "model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
426
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
427
+ "model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
428
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
429
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors",
430
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
431
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
432
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
433
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
434
+ "model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
435
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
436
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
437
+ "model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
438
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
439
+ "model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
440
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
441
+ "model.norm.weight": "model-00002-of-00002.safetensors"
442
+ }
443
+ }
Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
Blood/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/tokenizer_config.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
+ "errors": "replace",
201
+ "extra_special_tokens": {},
202
+ "model_max_length": 2048,
203
+ "pad_token": "<|endoftext|>",
204
+ "padding_side": "right",
205
+ "split_special_tokens": false,
206
+ "tokenizer_class": "Qwen2Tokenizer",
207
+ "unk_token": null
208
+ }
Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 11008,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention"
49
+ ],
50
+ "max_position_embeddings": 32768,
51
+ "max_window_layers": 70,
52
+ "model_type": "qwen2",
53
+ "num_attention_heads": 16,
54
+ "num_hidden_layers": 36,
55
+ "num_key_value_heads": 2,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.55.2",
63
+ "use_cache": true,
64
+ "use_sliding_window": false,
65
+ "vocab_size": 151936
66
+ }
Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.55.2"
14
+ }
Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/model.safetensors.index.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_parameters": 3085938688,
4
+ "total_size": 6794207232
5
+ },
6
+ "weight_map": {
7
+ "lm_head.weight": "model-00002-of-00002.safetensors",
8
+ "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
9
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
10
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
11
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
12
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
13
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
14
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
15
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
16
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
17
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
18
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
19
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
20
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
21
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors",
22
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
23
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
24
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
25
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
26
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
27
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
28
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
29
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
30
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
31
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
32
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
33
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors",
34
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
35
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
36
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
37
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
38
+ "model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
39
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
40
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
41
+ "model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
42
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
43
+ "model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
44
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
45
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors",
46
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
47
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
48
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
49
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
50
+ "model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
51
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
52
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
53
+ "model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
54
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
55
+ "model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
56
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
57
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors",
58
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
59
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
60
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
61
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
62
+ "model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
63
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
64
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
65
+ "model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
66
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
67
+ "model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
68
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
69
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
70
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
71
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
72
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
73
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
74
+ "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
75
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
76
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
77
+ "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
78
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
79
+ "model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
80
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
81
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
82
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
83
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
84
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
85
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
86
+ "model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
87
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
88
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
89
+ "model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
90
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
91
+ "model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
92
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
93
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors",
94
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
95
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
96
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
97
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
98
+ "model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
99
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
100
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
101
+ "model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
102
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
103
+ "model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
104
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
105
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors",
106
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
107
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
108
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
109
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
110
+ "model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
111
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
112
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
113
+ "model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
114
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
115
+ "model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
116
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
117
+ "model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors",
118
+ "model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
119
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
120
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
121
+ "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
122
+ "model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
123
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
124
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
125
+ "model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
126
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
127
+ "model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
128
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
129
+ "model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
130
+ "model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
131
+ "model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
132
+ "model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
133
+ "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
134
+ "model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
135
+ "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
136
+ "model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
137
+ "model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
138
+ "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
139
+ "model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
140
+ "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
141
+ "model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
142
+ "model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
143
+ "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
144
+ "model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
145
+ "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
146
+ "model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
147
+ "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
148
+ "model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
149
+ "model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
150
+ "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
151
+ "model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
152
+ "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
153
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
154
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
155
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
156
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
157
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
158
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
159
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
160
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
161
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
162
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
163
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
164
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
165
+ "model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
166
+ "model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
167
+ "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
168
+ "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
169
+ "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
170
+ "model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
171
+ "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
172
+ "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
173
+ "model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
174
+ "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
175
+ "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
176
+ "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
177
+ "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
178
+ "model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
179
+ "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
180
+ "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
181
+ "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
182
+ "model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
183
+ "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
184
+ "model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
185
+ "model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
186
+ "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
187
+ "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
188
+ "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
189
+ "model.layers.22.input_layernorm.weight": "model-00001-of-00002.safetensors",
190
+ "model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
191
+ "model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
192
+ "model.layers.22.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
193
+ "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
194
+ "model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
195
+ "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
196
+ "model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
197
+ "model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
198
+ "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
199
+ "model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
200
+ "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
201
+ "model.layers.23.input_layernorm.weight": "model-00001-of-00002.safetensors",
202
+ "model.layers.23.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
203
+ "model.layers.23.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
204
+ "model.layers.23.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
205
+ "model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
206
+ "model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
207
+ "model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
208
+ "model.layers.23.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
209
+ "model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
210
+ "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
211
+ "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
212
+ "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
213
+ "model.layers.24.input_layernorm.weight": "model-00001-of-00002.safetensors",
214
+ "model.layers.24.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
215
+ "model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
216
+ "model.layers.24.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
217
+ "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
218
+ "model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
219
+ "model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
220
+ "model.layers.24.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
221
+ "model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
222
+ "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
223
+ "model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
224
+ "model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
225
+ "model.layers.25.input_layernorm.weight": "model-00001-of-00002.safetensors",
226
+ "model.layers.25.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
227
+ "model.layers.25.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
228
+ "model.layers.25.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
229
+ "model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
230
+ "model.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
231
+ "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
232
+ "model.layers.25.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
233
+ "model.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
234
+ "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
235
+ "model.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
236
+ "model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
237
+ "model.layers.26.input_layernorm.weight": "model-00001-of-00002.safetensors",
238
+ "model.layers.26.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
239
+ "model.layers.26.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
240
+ "model.layers.26.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
241
+ "model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
242
+ "model.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
243
+ "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
244
+ "model.layers.26.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
245
+ "model.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
246
+ "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
247
+ "model.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
248
+ "model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
249
+ "model.layers.27.input_layernorm.weight": "model-00001-of-00002.safetensors",
250
+ "model.layers.27.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
251
+ "model.layers.27.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
252
+ "model.layers.27.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
253
+ "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
254
+ "model.layers.27.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
255
+ "model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
256
+ "model.layers.27.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
257
+ "model.layers.27.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
258
+ "model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
259
+ "model.layers.27.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
260
+ "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
261
+ "model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors",
262
+ "model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
263
+ "model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
264
+ "model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
265
+ "model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
266
+ "model.layers.28.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
267
+ "model.layers.28.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
268
+ "model.layers.28.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
269
+ "model.layers.28.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
270
+ "model.layers.28.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
271
+ "model.layers.28.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
272
+ "model.layers.28.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
273
+ "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
274
+ "model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
275
+ "model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
276
+ "model.layers.29.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
277
+ "model.layers.29.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
278
+ "model.layers.29.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
279
+ "model.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
280
+ "model.layers.29.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
281
+ "model.layers.29.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
282
+ "model.layers.29.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
283
+ "model.layers.29.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
284
+ "model.layers.29.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
285
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
286
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
287
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
288
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
289
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
290
+ "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
291
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
292
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
293
+ "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
294
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
295
+ "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
296
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
297
+ "model.layers.30.input_layernorm.weight": "model-00002-of-00002.safetensors",
298
+ "model.layers.30.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
299
+ "model.layers.30.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
300
+ "model.layers.30.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
301
+ "model.layers.30.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
302
+ "model.layers.30.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
303
+ "model.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
304
+ "model.layers.30.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
305
+ "model.layers.30.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
306
+ "model.layers.30.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
307
+ "model.layers.30.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
308
+ "model.layers.30.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
309
+ "model.layers.31.input_layernorm.weight": "model-00002-of-00002.safetensors",
310
+ "model.layers.31.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
311
+ "model.layers.31.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
312
+ "model.layers.31.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
313
+ "model.layers.31.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
314
+ "model.layers.31.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
315
+ "model.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
316
+ "model.layers.31.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
317
+ "model.layers.31.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
318
+ "model.layers.31.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
319
+ "model.layers.31.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
320
+ "model.layers.31.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
321
+ "model.layers.32.input_layernorm.weight": "model-00002-of-00002.safetensors",
322
+ "model.layers.32.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
323
+ "model.layers.32.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
324
+ "model.layers.32.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
325
+ "model.layers.32.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
326
+ "model.layers.32.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
327
+ "model.layers.32.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
328
+ "model.layers.32.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
329
+ "model.layers.32.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
330
+ "model.layers.32.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
331
+ "model.layers.32.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
332
+ "model.layers.32.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
333
+ "model.layers.33.input_layernorm.weight": "model-00002-of-00002.safetensors",
334
+ "model.layers.33.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
335
+ "model.layers.33.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
336
+ "model.layers.33.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
337
+ "model.layers.33.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
338
+ "model.layers.33.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
339
+ "model.layers.33.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
340
+ "model.layers.33.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
341
+ "model.layers.33.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
342
+ "model.layers.33.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
343
+ "model.layers.33.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
344
+ "model.layers.33.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
345
+ "model.layers.34.input_layernorm.weight": "model-00002-of-00002.safetensors",
346
+ "model.layers.34.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
347
+ "model.layers.34.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
348
+ "model.layers.34.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
349
+ "model.layers.34.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
350
+ "model.layers.34.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
351
+ "model.layers.34.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
352
+ "model.layers.34.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
353
+ "model.layers.34.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
354
+ "model.layers.34.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
355
+ "model.layers.34.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
356
+ "model.layers.34.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
357
+ "model.layers.35.input_layernorm.weight": "model-00002-of-00002.safetensors",
358
+ "model.layers.35.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
359
+ "model.layers.35.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
360
+ "model.layers.35.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
361
+ "model.layers.35.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
362
+ "model.layers.35.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
363
+ "model.layers.35.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
364
+ "model.layers.35.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
365
+ "model.layers.35.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
366
+ "model.layers.35.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
367
+ "model.layers.35.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
368
+ "model.layers.35.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
369
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
370
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
371
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
372
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
373
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
374
+ "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
375
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
376
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
377
+ "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
378
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
379
+ "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
380
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
381
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
382
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
383
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
384
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
385
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
386
+ "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
387
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
388
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
389
+ "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
390
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
391
+ "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
392
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
393
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
394
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
395
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
396
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
397
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
398
+ "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
399
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
400
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
401
+ "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
402
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
403
+ "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
404
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
405
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors",
406
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
407
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
408
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
409
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
410
+ "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
411
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
412
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
413
+ "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
414
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
415
+ "model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
416
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
417
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors",
418
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
419
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
420
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
421
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
422
+ "model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
423
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
424
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
425
+ "model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
426
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
427
+ "model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
428
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
429
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors",
430
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
431
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
432
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
433
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
434
+ "model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
435
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
436
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
437
+ "model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
438
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
439
+ "model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
440
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
441
+ "model.norm.weight": "model-00002-of-00002.safetensors"
442
+ }
443
+ }
Blood/seed_15/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/tokenizer_config.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
+ "errors": "replace",
201
+ "extra_special_tokens": {},
202
+ "model_max_length": 2048,
203
+ "pad_token": "<|endoftext|>",
204
+ "padding_side": "right",
205
+ "split_special_tokens": false,
206
+ "tokenizer_class": "Qwen2Tokenizer",
207
+ "unk_token": null
208
+ }
Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ model_name: LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized
4
+ tags:
5
+ - generated_from_trainer
6
+ - grpo
7
+ - trl
8
+ licence: license
9
+ ---
10
+
11
+ # Model Card for LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized
12
+
13
+ This model is a fine-tuned version of [None](https://huggingface.co/None).
14
+ It has been trained using [TRL](https://github.com/huggingface/trl).
15
+
16
+ ## Quick start
17
+
18
+ ```python
19
+ from transformers import pipeline
20
+
21
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
22
+ generator = pipeline("text-generation", model="None", device="cuda")
23
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
24
+ print(output["generated_text"])
25
+ ```
26
+
27
+ ## Training procedure
28
+
29
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/baileyyeah326/MLLM/runs/4xam9tu2)
30
+
31
+
32
+ This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
33
+
34
+ ### Framework versions
35
+
36
+ - TRL: 0.21.0
37
+ - Transformers: 4.55.2
38
+ - Pytorch: 2.4.0+cu124
39
+ - Datasets: 4.0.0
40
+ - Tokenizers: 0.21.4
41
+
42
+ ## Citations
43
+
44
+ Cite GRPO as:
45
+
46
+ ```bibtex
47
+ @article{zhihong2024deepseekmath,
48
+ title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
49
+ author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
50
+ year = 2024,
51
+ eprint = {arXiv:2402.03300},
52
+ }
53
+
54
+ ```
55
+
56
+ Cite TRL as:
57
+
58
+ ```bibtex
59
+ @misc{vonwerra2022trl,
60
+ title = {{TRL: Transformer Reinforcement Learning}},
61
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
62
+ year = 2020,
63
+ journal = {GitHub repository},
64
+ publisher = {GitHub},
65
+ howpublished = {\url{https://github.com/huggingface/trl}}
66
+ }
67
+ ```
Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 11008,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention"
49
+ ],
50
+ "max_position_embeddings": 32768,
51
+ "max_window_layers": 70,
52
+ "model_type": "qwen2",
53
+ "num_attention_heads": 16,
54
+ "num_hidden_layers": 36,
55
+ "num_key_value_heads": 2,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.55.2",
63
+ "use_cache": false,
64
+ "use_sliding_window": false,
65
+ "vocab_size": 151936
66
+ }
Blood/seed_2/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.55.2"
14
+ }
Blood/seed_2/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.55.2"
14
+ }
Blood/seed_2/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/model.safetensors.index.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_parameters": 3085938688,
4
+ "total_size": 6794207232
5
+ },
6
+ "weight_map": {
7
+ "lm_head.weight": "model-00002-of-00002.safetensors",
8
+ "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
9
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
10
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
11
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
12
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
13
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
14
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
15
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
16
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
17
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
18
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
19
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
20
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
21
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors",
22
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
23
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
24
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
25
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
26
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
27
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
28
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
29
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
30
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
31
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
32
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
33
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors",
34
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
35
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
36
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
37
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
38
+ "model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
39
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
40
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
41
+ "model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
42
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
43
+ "model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
44
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
45
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors",
46
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
47
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
48
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
49
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
50
+ "model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
51
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
52
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
53
+ "model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
54
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
55
+ "model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
56
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
57
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors",
58
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
59
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
60
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
61
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
62
+ "model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
63
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
64
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
65
+ "model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
66
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
67
+ "model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
68
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
69
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
70
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
71
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
72
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
73
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
74
+ "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
75
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
76
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
77
+ "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
78
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
79
+ "model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
80
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
81
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
82
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
83
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
84
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
85
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
86
+ "model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
87
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
88
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
89
+ "model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
90
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
91
+ "model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
92
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
93
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors",
94
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
95
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
96
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
97
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
98
+ "model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
99
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
100
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
101
+ "model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
102
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
103
+ "model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
104
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
105
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors",
106
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
107
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
108
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
109
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
110
+ "model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
111
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
112
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
113
+ "model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
114
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
115
+ "model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
116
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
117
+ "model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors",
118
+ "model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
119
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
120
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
121
+ "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
122
+ "model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
123
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
124
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
125
+ "model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
126
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
127
+ "model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
128
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
129
+ "model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
130
+ "model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
131
+ "model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
132
+ "model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
133
+ "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
134
+ "model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
135
+ "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
136
+ "model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
137
+ "model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
138
+ "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
139
+ "model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
140
+ "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
141
+ "model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
142
+ "model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
143
+ "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
144
+ "model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
145
+ "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
146
+ "model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
147
+ "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
148
+ "model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
149
+ "model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
150
+ "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
151
+ "model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
152
+ "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
153
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
154
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
155
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
156
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
157
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
158
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
159
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
160
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
161
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
162
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
163
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
164
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
165
+ "model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
166
+ "model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
167
+ "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
168
+ "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
169
+ "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
170
+ "model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
171
+ "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
172
+ "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
173
+ "model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
174
+ "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
175
+ "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
176
+ "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
177
+ "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
178
+ "model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
179
+ "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
180
+ "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
181
+ "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
182
+ "model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
183
+ "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
184
+ "model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
185
+ "model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
186
+ "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
187
+ "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
188
+ "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
189
+ "model.layers.22.input_layernorm.weight": "model-00001-of-00002.safetensors",
190
+ "model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
191
+ "model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
192
+ "model.layers.22.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
193
+ "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
194
+ "model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
195
+ "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
196
+ "model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
197
+ "model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
198
+ "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
199
+ "model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
200
+ "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
201
+ "model.layers.23.input_layernorm.weight": "model-00001-of-00002.safetensors",
202
+ "model.layers.23.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
203
+ "model.layers.23.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
204
+ "model.layers.23.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
205
+ "model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
206
+ "model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
207
+ "model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
208
+ "model.layers.23.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
209
+ "model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
210
+ "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
211
+ "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
212
+ "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
213
+ "model.layers.24.input_layernorm.weight": "model-00001-of-00002.safetensors",
214
+ "model.layers.24.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
215
+ "model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
216
+ "model.layers.24.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
217
+ "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
218
+ "model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
219
+ "model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
220
+ "model.layers.24.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
221
+ "model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
222
+ "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
223
+ "model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
224
+ "model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
225
+ "model.layers.25.input_layernorm.weight": "model-00001-of-00002.safetensors",
226
+ "model.layers.25.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
227
+ "model.layers.25.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
228
+ "model.layers.25.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
229
+ "model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
230
+ "model.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
231
+ "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
232
+ "model.layers.25.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
233
+ "model.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
234
+ "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
235
+ "model.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
236
+ "model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
237
+ "model.layers.26.input_layernorm.weight": "model-00001-of-00002.safetensors",
238
+ "model.layers.26.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
239
+ "model.layers.26.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
240
+ "model.layers.26.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
241
+ "model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
242
+ "model.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
243
+ "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
244
+ "model.layers.26.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
245
+ "model.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
246
+ "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
247
+ "model.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
248
+ "model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
249
+ "model.layers.27.input_layernorm.weight": "model-00001-of-00002.safetensors",
250
+ "model.layers.27.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
251
+ "model.layers.27.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
252
+ "model.layers.27.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
253
+ "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
254
+ "model.layers.27.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
255
+ "model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
256
+ "model.layers.27.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
257
+ "model.layers.27.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
258
+ "model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
259
+ "model.layers.27.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
260
+ "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
261
+ "model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors",
262
+ "model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
263
+ "model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
264
+ "model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
265
+ "model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
266
+ "model.layers.28.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
267
+ "model.layers.28.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
268
+ "model.layers.28.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
269
+ "model.layers.28.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
270
+ "model.layers.28.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
271
+ "model.layers.28.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
272
+ "model.layers.28.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
273
+ "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
274
+ "model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
275
+ "model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
276
+ "model.layers.29.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
277
+ "model.layers.29.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
278
+ "model.layers.29.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
279
+ "model.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
280
+ "model.layers.29.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
281
+ "model.layers.29.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
282
+ "model.layers.29.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
283
+ "model.layers.29.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
284
+ "model.layers.29.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
285
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
286
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
287
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
288
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
289
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
290
+ "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
291
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
292
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
293
+ "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
294
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
295
+ "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
296
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
297
+ "model.layers.30.input_layernorm.weight": "model-00002-of-00002.safetensors",
298
+ "model.layers.30.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
299
+ "model.layers.30.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
300
+ "model.layers.30.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
301
+ "model.layers.30.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
302
+ "model.layers.30.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
303
+ "model.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
304
+ "model.layers.30.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
305
+ "model.layers.30.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
306
+ "model.layers.30.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
307
+ "model.layers.30.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
308
+ "model.layers.30.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
309
+ "model.layers.31.input_layernorm.weight": "model-00002-of-00002.safetensors",
310
+ "model.layers.31.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
311
+ "model.layers.31.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
312
+ "model.layers.31.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
313
+ "model.layers.31.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
314
+ "model.layers.31.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
315
+ "model.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
316
+ "model.layers.31.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
317
+ "model.layers.31.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
318
+ "model.layers.31.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
319
+ "model.layers.31.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
320
+ "model.layers.31.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
321
+ "model.layers.32.input_layernorm.weight": "model-00002-of-00002.safetensors",
322
+ "model.layers.32.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
323
+ "model.layers.32.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
324
+ "model.layers.32.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
325
+ "model.layers.32.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
326
+ "model.layers.32.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
327
+ "model.layers.32.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
328
+ "model.layers.32.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
329
+ "model.layers.32.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
330
+ "model.layers.32.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
331
+ "model.layers.32.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
332
+ "model.layers.32.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
333
+ "model.layers.33.input_layernorm.weight": "model-00002-of-00002.safetensors",
334
+ "model.layers.33.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
335
+ "model.layers.33.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
336
+ "model.layers.33.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
337
+ "model.layers.33.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
338
+ "model.layers.33.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
339
+ "model.layers.33.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
340
+ "model.layers.33.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
341
+ "model.layers.33.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
342
+ "model.layers.33.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
343
+ "model.layers.33.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
344
+ "model.layers.33.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
345
+ "model.layers.34.input_layernorm.weight": "model-00002-of-00002.safetensors",
346
+ "model.layers.34.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
347
+ "model.layers.34.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
348
+ "model.layers.34.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
349
+ "model.layers.34.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
350
+ "model.layers.34.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
351
+ "model.layers.34.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
352
+ "model.layers.34.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
353
+ "model.layers.34.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
354
+ "model.layers.34.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
355
+ "model.layers.34.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
356
+ "model.layers.34.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
357
+ "model.layers.35.input_layernorm.weight": "model-00002-of-00002.safetensors",
358
+ "model.layers.35.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
359
+ "model.layers.35.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
360
+ "model.layers.35.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
361
+ "model.layers.35.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
362
+ "model.layers.35.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
363
+ "model.layers.35.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
364
+ "model.layers.35.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
365
+ "model.layers.35.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
366
+ "model.layers.35.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
367
+ "model.layers.35.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
368
+ "model.layers.35.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
369
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
370
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
371
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
372
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
373
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
374
+ "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
375
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
376
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
377
+ "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
378
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
379
+ "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
380
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
381
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
382
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
383
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
384
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
385
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
386
+ "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
387
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
388
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
389
+ "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
390
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
391
+ "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
392
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
393
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
394
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
395
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
396
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
397
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
398
+ "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
399
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
400
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
401
+ "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
402
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
403
+ "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
404
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
405
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors",
406
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
407
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
408
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
409
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
410
+ "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
411
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
412
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
413
+ "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
414
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
415
+ "model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
416
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
417
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors",
418
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
419
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
420
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
421
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
422
+ "model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
423
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
424
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
425
+ "model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
426
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
427
+ "model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
428
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
429
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors",
430
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
431
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
432
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
433
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
434
+ "model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
435
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
436
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
437
+ "model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
438
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
439
+ "model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
440
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
441
+ "model.norm.weight": "model-00002-of-00002.safetensors"
442
+ }
443
+ }
Blood/seed_2/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
Blood/seed_2/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/tokenizer_config.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
+ "errors": "replace",
201
+ "extra_special_tokens": {},
202
+ "model_max_length": 2048,
203
+ "pad_token": "<|endoftext|>",
204
+ "padding_side": "right",
205
+ "split_special_tokens": false,
206
+ "tokenizer_class": "Qwen2Tokenizer",
207
+ "unk_token": null
208
+ }
Blood/seed_4/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ model_name: LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized
4
+ tags:
5
+ - generated_from_trainer
6
+ - grpo
7
+ - trl
8
+ licence: license
9
+ ---
10
+
11
+ # Model Card for LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized
12
+
13
+ This model is a fine-tuned version of [None](https://huggingface.co/None).
14
+ It has been trained using [TRL](https://github.com/huggingface/trl).
15
+
16
+ ## Quick start
17
+
18
+ ```python
19
+ from transformers import pipeline
20
+
21
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
22
+ generator = pipeline("text-generation", model="None", device="cuda")
23
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
24
+ print(output["generated_text"])
25
+ ```
26
+
27
+ ## Training procedure
28
+
29
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/baileyyeah326/MLLM/runs/vdh21tyh)
30
+
31
+
32
+ This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
33
+
34
+ ### Framework versions
35
+
36
+ - TRL: 0.21.0
37
+ - Transformers: 4.55.2
38
+ - Pytorch: 2.4.0+cu124
39
+ - Datasets: 4.0.0
40
+ - Tokenizers: 0.21.4
41
+
42
+ ## Citations
43
+
44
+ Cite GRPO as:
45
+
46
+ ```bibtex
47
+ @article{zhihong2024deepseekmath,
48
+ title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
49
+ author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
50
+ year = 2024,
51
+ eprint = {arXiv:2402.03300},
52
+ }
53
+
54
+ ```
55
+
56
+ Cite TRL as:
57
+
58
+ ```bibtex
59
+ @misc{vonwerra2022trl,
60
+ title = {{TRL: Transformer Reinforcement Learning}},
61
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
62
+ year = 2020,
63
+ journal = {GitHub repository},
64
+ publisher = {GitHub},
65
+ howpublished = {\url{https://github.com/huggingface/trl}}
66
+ }
67
+ ```
Blood/seed_4/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Blood/seed_4/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Blood/seed_4/GRPO/LLM-Qwen-2.5-3B-GRPO-decision-tree-Blood-serialized/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 11008,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention"
49
+ ],
50
+ "max_position_embeddings": 32768,
51
+ "max_window_layers": 70,
52
+ "model_type": "qwen2",
53
+ "num_attention_heads": 16,
54
+ "num_hidden_layers": 36,
55
+ "num_key_value_heads": 2,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.55.2",
63
+ "use_cache": false,
64
+ "use_sliding_window": false,
65
+ "vocab_size": 151936
66
+ }
Blood/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Blood/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Blood/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 11008,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention"
49
+ ],
50
+ "max_position_embeddings": 32768,
51
+ "max_window_layers": 70,
52
+ "model_type": "qwen2",
53
+ "num_attention_heads": 16,
54
+ "num_hidden_layers": 36,
55
+ "num_key_value_heads": 2,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.55.2",
63
+ "use_cache": true,
64
+ "use_sliding_window": false,
65
+ "vocab_size": 151936
66
+ }
Blood/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.55.2"
14
+ }
Blood/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Blood/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Blood/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 11008,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention"
49
+ ],
50
+ "max_position_embeddings": 32768,
51
+ "max_window_layers": 70,
52
+ "model_type": "qwen2",
53
+ "num_attention_heads": 16,
54
+ "num_hidden_layers": 36,
55
+ "num_key_value_heads": 2,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.55.2",
63
+ "use_cache": true,
64
+ "use_sliding_window": false,
65
+ "vocab_size": 151936
66
+ }
Blood/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Blood-serialized/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.55.2"
14
+ }
Creditg/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Creditg/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Creditg/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 11008,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention"
49
+ ],
50
+ "max_position_embeddings": 32768,
51
+ "max_window_layers": 70,
52
+ "model_type": "qwen2",
53
+ "num_attention_heads": 16,
54
+ "num_hidden_layers": 36,
55
+ "num_key_value_heads": 2,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.55.2",
63
+ "use_cache": true,
64
+ "use_sliding_window": false,
65
+ "vocab_size": 151936
66
+ }
Creditg/seed_0/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.55.2"
14
+ }
Creditg/seed_4/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/runs/Dec26_23-58-50_csce-yang-s2.engr.tamu.edu/events.out.tfevents.1766815143.csce-yang-s2.engr.tamu.edu.3648358.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63c500daef15e1649376263aaec8c28e14ea7b9b2668053e9a12f3242015fbb8
3
+ size 206473
Creditg/seed_50/SFT/LLM-Qwen-2.5-3B-SFT-decision-tree-Creditg-serialized/runs/Dec27_07-28-09_csce-yang-s2.engr.tamu.edu/events.out.tfevents.1766842103.csce-yang-s2.engr.tamu.edu.3873179.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5514891fbf86d3a176725ac2df323f71ed0600e70d35faee1355408455d0c05
3
+ size 186280