antoine-444 commited on
Commit
bc047a0
·
verified ·
1 Parent(s): e51839d

Upload folder using huggingface_hub

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:62a9621d627d4b69570f9471e368771d7548798ba88f3c0eb317ce150d9a3d70
3
  size 1192135096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d98edc3d2ca1aec91433e9987eee768c36fddb0d2cde139c9b53415ce830ce40
3
  size 1192135096
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b535e39d606564233a120331ad9e0121c45ae5b345ddc807d5c8a7366d4e1002
3
  size 2384459962
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9069b63adcfe2746e98414c2c83d0ddd2f1f9d6d20a61a295651cba33be78589
3
  size 2384459962
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:386fcc8cc1089aade9450d86fb239ea3483f455fd2d78d8378645feecfec9d69
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d245e05e72192c132e0f2edb6fdcae0c578c890f0fe912f17ec7b0bba2d38cc3
3
  size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf649bac0f2cfebbcf7fed3fb87fad328f4aea97a5892dffe764f4191f92f88c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3062b0a02968eef37b0d0a0387798ef4cd562c0a577e5e8055c6181ccd441515
3
  size 1064
special_tokens_map.json CHANGED
@@ -15,11 +15,11 @@
15
  "<|video_pad|>"
16
  ],
17
  "eos_token": {
18
- "content": "<|im_end|>",
19
  "lstrip": false,
20
  "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
- "pad_token": "<|im_end|>"
25
  }
 
15
  "<|video_pad|>"
16
  ],
17
  "eos_token": {
18
+ "content": "<|endoftext|>",
19
  "lstrip": false,
20
  "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
+ "pad_token": "<|endoftext|>"
25
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
- size 11422654
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:574de68a0f63f2004784a421c7d42c2b2786c05cb38542d2ed3525757a1f7fde
3
+ size 11422932
tokenizer_config.json CHANGED
@@ -229,11 +229,11 @@
229
  "bos_token": null,
230
  "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0].role == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0].content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}\n{%- for message in messages[::-1] %}\n {%- set index = (messages|length - 1) - loop.index0 %}\n {%- if ns.multi_step_tool and message.role == \"user\" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}\n {%- set ns.multi_step_tool = false %}\n {%- set ns.last_query_index = index %}\n {%- endif %}\n{%- endfor %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {%- set content = message.content %}\n {%- set reasoning_content = '' %}\n {%- if message.reasoning_content is defined and message.reasoning_content is not none %}\n {%- set reasoning_content = message.reasoning_content %}\n {%- else %}\n {%- if '</think>' in message.content %}\n {%- set content = message.content.split('</think>')[-1].lstrip('\\n') %}\n {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\\n').split('<think>')[-1].lstrip('\\n') %}\n {%- endif %}\n {%- endif %}\n {%- if loop.index0 > ns.last_query_index %}\n {%- if loop.last or (not loop.last and reasoning_content) %}\n {{- '<|im_start|>' + message.role + '\\n<think>\\n' + reasoning_content.strip('\\n') + '\\n</think>\\n\\n' + content.lstrip('\\n') }}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- if message.tool_calls %}\n {%- for tool_call in message.tool_calls %}\n {%- if (loop.first and content) or (not loop.first) %}\n {{- '\\n' }}\n {%- endif %}\n {%- if tool_call.function %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {%- if tool_call.arguments is string %}\n {{- tool_call.arguments }}\n {%- else %}\n {{- tool_call.arguments | tojson }}\n {%- endif %}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {%- endif %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if loop.first or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n {%- if enable_thinking is defined and enable_thinking is false %}\n {{- '<think>\\n\\n</think>\\n\\n' }}\n {%- endif %}\n{%- endif %}",
231
  "clean_up_tokenization_spaces": false,
232
- "eos_token": "<|im_end|>",
233
  "errors": "replace",
234
  "extra_special_tokens": {},
235
  "model_max_length": 131072,
236
- "pad_token": "<|im_end|>",
237
  "split_special_tokens": false,
238
  "tokenizer_class": "Qwen2Tokenizer",
239
  "unk_token": null
 
229
  "bos_token": null,
230
  "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0].role == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0].content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}\n{%- for message in messages[::-1] %}\n {%- set index = (messages|length - 1) - loop.index0 %}\n {%- if ns.multi_step_tool and message.role == \"user\" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}\n {%- set ns.multi_step_tool = false %}\n {%- set ns.last_query_index = index %}\n {%- endif %}\n{%- endfor %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {%- set content = message.content %}\n {%- set reasoning_content = '' %}\n {%- if message.reasoning_content is defined and message.reasoning_content is not none %}\n {%- set reasoning_content = message.reasoning_content %}\n {%- else %}\n {%- if '</think>' in message.content %}\n {%- set content = message.content.split('</think>')[-1].lstrip('\\n') %}\n {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\\n').split('<think>')[-1].lstrip('\\n') %}\n {%- endif %}\n {%- endif %}\n {%- if loop.index0 > ns.last_query_index %}\n {%- if loop.last or (not loop.last and reasoning_content) %}\n {{- '<|im_start|>' + message.role + '\\n<think>\\n' + reasoning_content.strip('\\n') + '\\n</think>\\n\\n' + content.lstrip('\\n') }}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- if message.tool_calls %}\n {%- for tool_call in message.tool_calls %}\n {%- if (loop.first and content) or (not loop.first) %}\n {{- '\\n' }}\n {%- endif %}\n {%- if tool_call.function %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {%- if tool_call.arguments is string %}\n {{- tool_call.arguments }}\n {%- else %}\n {{- tool_call.arguments | tojson }}\n {%- endif %}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {%- endif %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if loop.first or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n {%- if enable_thinking is defined and enable_thinking is false %}\n {{- '<think>\\n\\n</think>\\n\\n' }}\n {%- endif %}\n{%- endif %}",
231
  "clean_up_tokenization_spaces": false,
232
+ "eos_token": "<|endoftext|>",
233
  "errors": "replace",
234
  "extra_special_tokens": {},
235
  "model_max_length": 131072,
236
+ "pad_token": "<|endoftext|>",
237
  "split_special_tokens": false,
238
  "tokenizer_class": "Qwen2Tokenizer",
239
  "unk_token": null
trainer_state.json CHANGED
@@ -2,100 +2,151 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.9984,
6
- "eval_steps": 500,
7
- "global_step": 312,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "epoch": 0.08,
14
- "grad_norm": 6.90625,
15
- "learning_rate": 4.615384615384616e-05,
16
- "loss": 0.9872,
17
- "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  },
19
  {
20
- "epoch": 0.16,
21
- "grad_norm": 7.0,
22
- "learning_rate": 4.21474358974359e-05,
23
- "loss": 0.9209,
24
- "step": 50
 
25
  },
26
  {
27
- "epoch": 0.24,
28
- "grad_norm": 6.90625,
29
- "learning_rate": 3.814102564102564e-05,
30
- "loss": 0.866,
31
- "step": 75
32
  },
33
  {
34
  "epoch": 0.32,
35
- "grad_norm": 6.09375,
36
- "learning_rate": 3.4134615384615386e-05,
37
- "loss": 0.8653,
38
- "step": 100
 
39
  },
40
  {
41
- "epoch": 0.4,
42
- "grad_norm": 7.09375,
43
- "learning_rate": 3.012820512820513e-05,
44
- "loss": 0.8149,
45
- "step": 125
46
  },
47
  {
48
- "epoch": 0.48,
49
- "grad_norm": 7.40625,
50
- "learning_rate": 2.6121794871794876e-05,
51
- "loss": 0.8612,
52
- "step": 150
 
53
  },
54
  {
55
- "epoch": 0.56,
56
- "grad_norm": 7.3125,
57
- "learning_rate": 2.2115384615384616e-05,
58
- "loss": 0.8275,
59
- "step": 175
 
 
 
 
 
 
 
 
60
  },
61
  {
62
  "epoch": 0.64,
63
- "grad_norm": 7.0625,
64
- "learning_rate": 1.810897435897436e-05,
65
- "loss": 0.8186,
66
- "step": 200
 
 
 
 
 
 
 
 
67
  },
68
  {
69
- "epoch": 0.72,
70
- "grad_norm": 6.96875,
71
- "learning_rate": 1.4102564102564104e-05,
72
- "loss": 0.8048,
73
- "step": 225
74
  },
75
  {
76
- "epoch": 0.8,
77
- "grad_norm": 6.625,
78
- "learning_rate": 1.0096153846153847e-05,
79
- "loss": 0.8237,
80
- "step": 250
 
81
  },
82
  {
83
- "epoch": 0.88,
84
- "grad_norm": 7.75,
85
- "learning_rate": 6.08974358974359e-06,
86
- "loss": 0.8544,
87
- "step": 275
 
 
 
 
 
 
 
 
88
  },
89
  {
90
  "epoch": 0.96,
91
- "grad_norm": 7.15625,
92
- "learning_rate": 2.0833333333333334e-06,
93
- "loss": 0.8185,
94
- "step": 300
 
 
 
 
 
 
 
 
95
  }
96
  ],
97
- "logging_steps": 25,
98
- "max_steps": 312,
99
  "num_input_tokens_seen": 0,
100
  "num_train_epochs": 1,
101
  "save_steps": 500,
@@ -111,7 +162,7 @@
111
  "attributes": {}
112
  }
113
  },
114
- "total_flos": 3377377597980672.0,
115
  "train_batch_size": 4,
116
  "trial_name": null,
117
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.9994666666666666,
6
+ "eval_steps": 100,
7
+ "global_step": 937,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "epoch": 0.10666666666666667,
14
+ "grad_norm": 6.6875,
15
+ "learning_rate": 4.471718249733191e-05,
16
+ "loss": 1.2181,
17
+ "step": 100
18
+ },
19
+ {
20
+ "epoch": 0.10666666666666667,
21
+ "eval_loss": 1.4656692743301392,
22
+ "eval_runtime": 128.3122,
23
+ "eval_samples_per_second": 44.703,
24
+ "eval_steps_per_second": 5.588,
25
+ "step": 100
26
+ },
27
+ {
28
+ "epoch": 0.21333333333333335,
29
+ "grad_norm": 6.5625,
30
+ "learning_rate": 3.938100320170758e-05,
31
+ "loss": 1.1539,
32
+ "step": 200
33
  },
34
  {
35
+ "epoch": 0.21333333333333335,
36
+ "eval_loss": 1.4585195779800415,
37
+ "eval_runtime": 128.2644,
38
+ "eval_samples_per_second": 44.72,
39
+ "eval_steps_per_second": 5.59,
40
+ "step": 200
41
  },
42
  {
43
+ "epoch": 0.32,
44
+ "grad_norm": 5.625,
45
+ "learning_rate": 3.4044823906083245e-05,
46
+ "loss": 1.1353,
47
+ "step": 300
48
  },
49
  {
50
  "epoch": 0.32,
51
+ "eval_loss": 1.4478052854537964,
52
+ "eval_runtime": 128.1523,
53
+ "eval_samples_per_second": 44.759,
54
+ "eval_steps_per_second": 5.595,
55
+ "step": 300
56
  },
57
  {
58
+ "epoch": 0.4266666666666667,
59
+ "grad_norm": 5.8125,
60
+ "learning_rate": 2.8708644610458914e-05,
61
+ "loss": 1.133,
62
+ "step": 400
63
  },
64
  {
65
+ "epoch": 0.4266666666666667,
66
+ "eval_loss": 1.4380828142166138,
67
+ "eval_runtime": 128.2602,
68
+ "eval_samples_per_second": 44.722,
69
+ "eval_steps_per_second": 5.59,
70
+ "step": 400
71
  },
72
  {
73
+ "epoch": 0.5333333333333333,
74
+ "grad_norm": 5.4375,
75
+ "learning_rate": 2.337246531483458e-05,
76
+ "loss": 1.1229,
77
+ "step": 500
78
+ },
79
+ {
80
+ "epoch": 0.5333333333333333,
81
+ "eval_loss": 1.4234306812286377,
82
+ "eval_runtime": 128.3478,
83
+ "eval_samples_per_second": 44.691,
84
+ "eval_steps_per_second": 5.586,
85
+ "step": 500
86
  },
87
  {
88
  "epoch": 0.64,
89
+ "grad_norm": 5.71875,
90
+ "learning_rate": 1.8036286019210247e-05,
91
+ "loss": 1.1073,
92
+ "step": 600
93
+ },
94
+ {
95
+ "epoch": 0.64,
96
+ "eval_loss": 1.4192696809768677,
97
+ "eval_runtime": 128.1918,
98
+ "eval_samples_per_second": 44.745,
99
+ "eval_steps_per_second": 5.593,
100
+ "step": 600
101
  },
102
  {
103
+ "epoch": 0.7466666666666667,
104
+ "grad_norm": 6.4375,
105
+ "learning_rate": 1.2700106723585914e-05,
106
+ "loss": 1.1053,
107
+ "step": 700
108
  },
109
  {
110
+ "epoch": 0.7466666666666667,
111
+ "eval_loss": 1.4158320426940918,
112
+ "eval_runtime": 128.2181,
113
+ "eval_samples_per_second": 44.736,
114
+ "eval_steps_per_second": 5.592,
115
+ "step": 700
116
  },
117
  {
118
+ "epoch": 0.8533333333333334,
119
+ "grad_norm": 6.03125,
120
+ "learning_rate": 7.363927427961579e-06,
121
+ "loss": 1.1059,
122
+ "step": 800
123
+ },
124
+ {
125
+ "epoch": 0.8533333333333334,
126
+ "eval_loss": 1.4143309593200684,
127
+ "eval_runtime": 128.2757,
128
+ "eval_samples_per_second": 44.716,
129
+ "eval_steps_per_second": 5.59,
130
+ "step": 800
131
  },
132
  {
133
  "epoch": 0.96,
134
+ "grad_norm": 5.75,
135
+ "learning_rate": 2.0277481323372464e-06,
136
+ "loss": 1.1033,
137
+ "step": 900
138
+ },
139
+ {
140
+ "epoch": 0.96,
141
+ "eval_loss": 1.413725733757019,
142
+ "eval_runtime": 128.2264,
143
+ "eval_samples_per_second": 44.733,
144
+ "eval_steps_per_second": 5.592,
145
+ "step": 900
146
  }
147
  ],
148
+ "logging_steps": 100,
149
+ "max_steps": 937,
150
  "num_input_tokens_seen": 0,
151
  "num_train_epochs": 1,
152
  "save_steps": 500,
 
162
  "attributes": {}
163
  }
164
  },
165
+ "total_flos": 2.0285915444281344e+16,
166
  "train_batch_size": 4,
167
  "trial_name": null,
168
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fdc44085ea5aec9b20c61210d797707f5d128a458b13bba0602952888ce6f88d
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e22a09f7d57eefb94c0225ca0ecccdde6d19aff2eb73e7914dd16e4c7c4bc5f
3
  size 5304