skdrx commited on
Commit
0e727a9
·
verified ·
1 Parent(s): 3dd5395

Initial release of Dracula Flow base model

Browse files
Files changed (44) hide show
  1. .gitattributes +3 -0
  2. .ipynb_checkpoints/README-checkpoint.md +58 -0
  3. README.md +58 -0
  4. added_tokens.json +28 -0
  5. chat_template.jinja +85 -0
  6. checkpoint-2300/added_tokens.json +28 -0
  7. checkpoint-2300/chat_template.jinja +85 -0
  8. checkpoint-2300/config.json +60 -0
  9. checkpoint-2300/generation_config.json +8 -0
  10. checkpoint-2300/merges.txt +0 -0
  11. checkpoint-2300/model.safetensors +3 -0
  12. checkpoint-2300/optimizer.pt +3 -0
  13. checkpoint-2300/rng_state.pth +3 -0
  14. checkpoint-2300/scheduler.pt +3 -0
  15. checkpoint-2300/special_tokens_map.json +31 -0
  16. checkpoint-2300/tokenizer.json +3 -0
  17. checkpoint-2300/tokenizer_config.json +239 -0
  18. checkpoint-2300/trainer_state.json +2334 -0
  19. checkpoint-2300/training_args.bin +3 -0
  20. checkpoint-2300/vocab.json +0 -0
  21. checkpoint-2328/added_tokens.json +28 -0
  22. checkpoint-2328/chat_template.jinja +85 -0
  23. checkpoint-2328/config.json +60 -0
  24. checkpoint-2328/generation_config.json +8 -0
  25. checkpoint-2328/merges.txt +0 -0
  26. checkpoint-2328/model.safetensors +3 -0
  27. checkpoint-2328/optimizer.pt +3 -0
  28. checkpoint-2328/rng_state.pth +3 -0
  29. checkpoint-2328/scheduler.pt +3 -0
  30. checkpoint-2328/special_tokens_map.json +31 -0
  31. checkpoint-2328/tokenizer.json +3 -0
  32. checkpoint-2328/tokenizer_config.json +239 -0
  33. checkpoint-2328/trainer_state.json +2354 -0
  34. checkpoint-2328/training_args.bin +3 -0
  35. checkpoint-2328/vocab.json +0 -0
  36. config.json +60 -0
  37. generation_config.json +8 -0
  38. merges.txt +0 -0
  39. model.safetensors +3 -0
  40. special_tokens_map.json +31 -0
  41. tokenizer.json +3 -0
  42. tokenizer_config.json +239 -0
  43. training_args.bin +3 -0
  44. vocab.json +0 -0
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint-2300/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ checkpoint-2328/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
.ipynb_checkpoints/README-checkpoint.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-1.7B-Base
3
+ library_name: transformers
4
+ model_name: qwen3-sft-output
5
+ tags:
6
+ - generated_from_trainer
7
+ - sft
8
+ - trl
9
+ licence: license
10
+ ---
11
+
12
+ # Model Card for qwen3-sft-output
13
+
14
+ This model is a fine-tuned version of [Qwen/Qwen3-1.7B-Base](https://huggingface.co/Qwen/Qwen3-1.7B-Base).
15
+ It has been trained using [TRL](https://github.com/huggingface/trl).
16
+
17
+ ## Quick start
18
+
19
+ ```python
20
+ from transformers import pipeline
21
+
22
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="None", device="cuda")
24
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
+ print(output["generated_text"])
26
+ ```
27
+
28
+ ## Training procedure
29
+
30
+
31
+
32
+
33
+ This model was trained with SFT.
34
+
35
+ ### Framework versions
36
+
37
+ - TRL: 0.26.2
38
+ - Transformers: 4.57.3
39
+ - Pytorch: 2.7.0
40
+ - Datasets: 4.4.2
41
+ - Tokenizers: 0.22.2
42
+
43
+ ## Citations
44
+
45
+
46
+
47
+ Cite TRL as:
48
+
49
+ ```bibtex
50
+ @misc{vonwerra2022trl,
51
+ title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
+ year = 2020,
54
+ journal = {GitHub repository},
55
+ publisher = {GitHub},
56
+ howpublished = {\url{https://github.com/huggingface/trl}}
57
+ }
58
+ ```
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-1.7B-Base
3
+ library_name: transformers
4
+ model_name: qwen3-sft-output
5
+ tags:
6
+ - generated_from_trainer
7
+ - sft
8
+ - trl
9
+ licence: license
10
+ ---
11
+
12
+ # Model Card for qwen3-sft-output
13
+
14
+ This model is a fine-tuned version of [Qwen/Qwen3-1.7B-Base](https://huggingface.co/Qwen/Qwen3-1.7B-Base).
15
+ It has been trained using [TRL](https://github.com/huggingface/trl).
16
+
17
+ ## Quick start
18
+
19
+ ```python
20
+ from transformers import pipeline
21
+
22
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="None", device="cuda")
24
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
+ print(output["generated_text"])
26
+ ```
27
+
28
+ ## Training procedure
29
+
30
+
31
+
32
+
33
+ This model was trained with SFT.
34
+
35
+ ### Framework versions
36
+
37
+ - TRL: 0.26.2
38
+ - Transformers: 4.57.3
39
+ - Pytorch: 2.7.0
40
+ - Datasets: 4.4.2
41
+ - Tokenizers: 0.22.2
42
+
43
+ ## Citations
44
+
45
+
46
+
47
+ Cite TRL as:
48
+
49
+ ```bibtex
50
+ @misc{vonwerra2022trl,
51
+ title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
+ year = 2020,
54
+ journal = {GitHub repository},
55
+ publisher = {GitHub},
56
+ howpublished = {\url{https://github.com/huggingface/trl}}
57
+ }
58
+ ```
added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
chat_template.jinja ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
27
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
28
+ {%- elif message.role == "assistant" %}
29
+ {%- set content = message.content %}
30
+ {%- set reasoning_content = '' %}
31
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
32
+ {%- set reasoning_content = message.reasoning_content %}
33
+ {%- else %}
34
+ {%- if '</think>' in message.content %}
35
+ {%- set content = message.content.split('</think>')[-1].lstrip('\n') %}
36
+ {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
37
+ {%- endif %}
38
+ {%- endif %}
39
+ {%- if loop.index0 > ns.last_query_index %}
40
+ {%- if loop.last or (not loop.last and reasoning_content) %}
41
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
42
+ {%- else %}
43
+ {{- '<|im_start|>' + message.role + '\n' + content }}
44
+ {%- endif %}
45
+ {%- else %}
46
+ {{- '<|im_start|>' + message.role + '\n' + content }}
47
+ {%- endif %}
48
+ {%- if message.tool_calls %}
49
+ {%- for tool_call in message.tool_calls %}
50
+ {%- if (loop.first and content) or (not loop.first) %}
51
+ {{- '\n' }}
52
+ {%- endif %}
53
+ {%- if tool_call.function %}
54
+ {%- set tool_call = tool_call.function %}
55
+ {%- endif %}
56
+ {{- '<tool_call>\n{"name": "' }}
57
+ {{- tool_call.name }}
58
+ {{- '", "arguments": ' }}
59
+ {%- if tool_call.arguments is string %}
60
+ {{- tool_call.arguments }}
61
+ {%- else %}
62
+ {{- tool_call.arguments | tojson }}
63
+ {%- endif %}
64
+ {{- '}\n</tool_call>' }}
65
+ {%- endfor %}
66
+ {%- endif %}
67
+ {{- '<|im_end|>\n' }}
68
+ {%- elif message.role == "tool" %}
69
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
70
+ {{- '<|im_start|>user' }}
71
+ {%- endif %}
72
+ {{- '\n<tool_response>\n' }}
73
+ {{- message.content }}
74
+ {{- '\n</tool_response>' }}
75
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
76
+ {{- '<|im_end|>\n' }}
77
+ {%- endif %}
78
+ {%- endif %}
79
+ {%- endfor %}
80
+ {%- if add_generation_prompt %}
81
+ {{- '<|im_start|>assistant\n' }}
82
+ {%- if enable_thinking is defined and enable_thinking is false %}
83
+ {{- '<think>\n\n</think>\n\n' }}
84
+ {%- endif %}
85
+ {%- endif %}
checkpoint-2300/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
checkpoint-2300/chat_template.jinja ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
27
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
28
+ {%- elif message.role == "assistant" %}
29
+ {%- set content = message.content %}
30
+ {%- set reasoning_content = '' %}
31
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
32
+ {%- set reasoning_content = message.reasoning_content %}
33
+ {%- else %}
34
+ {%- if '</think>' in message.content %}
35
+ {%- set content = message.content.split('</think>')[-1].lstrip('\n') %}
36
+ {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
37
+ {%- endif %}
38
+ {%- endif %}
39
+ {%- if loop.index0 > ns.last_query_index %}
40
+ {%- if loop.last or (not loop.last and reasoning_content) %}
41
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
42
+ {%- else %}
43
+ {{- '<|im_start|>' + message.role + '\n' + content }}
44
+ {%- endif %}
45
+ {%- else %}
46
+ {{- '<|im_start|>' + message.role + '\n' + content }}
47
+ {%- endif %}
48
+ {%- if message.tool_calls %}
49
+ {%- for tool_call in message.tool_calls %}
50
+ {%- if (loop.first and content) or (not loop.first) %}
51
+ {{- '\n' }}
52
+ {%- endif %}
53
+ {%- if tool_call.function %}
54
+ {%- set tool_call = tool_call.function %}
55
+ {%- endif %}
56
+ {{- '<tool_call>\n{"name": "' }}
57
+ {{- tool_call.name }}
58
+ {{- '", "arguments": ' }}
59
+ {%- if tool_call.arguments is string %}
60
+ {{- tool_call.arguments }}
61
+ {%- else %}
62
+ {{- tool_call.arguments | tojson }}
63
+ {%- endif %}
64
+ {{- '}\n</tool_call>' }}
65
+ {%- endfor %}
66
+ {%- endif %}
67
+ {{- '<|im_end|>\n' }}
68
+ {%- elif message.role == "tool" %}
69
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
70
+ {{- '<|im_start|>user' }}
71
+ {%- endif %}
72
+ {{- '\n<tool_response>\n' }}
73
+ {{- message.content }}
74
+ {{- '\n</tool_response>' }}
75
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
76
+ {{- '<|im_end|>\n' }}
77
+ {%- endif %}
78
+ {%- endif %}
79
+ {%- endfor %}
80
+ {%- if add_generation_prompt %}
81
+ {{- '<|im_start|>assistant\n' }}
82
+ {%- if enable_thinking is defined and enable_thinking is false %}
83
+ {{- '<think>\n\n</think>\n\n' }}
84
+ {%- endif %}
85
+ {%- endif %}
checkpoint-2300/config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "dtype": "bfloat16",
8
+ "eos_token_id": 151643,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 2048,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 6144,
14
+ "layer_types": [
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention"
43
+ ],
44
+ "max_position_embeddings": 32768,
45
+ "max_window_layers": 28,
46
+ "model_type": "qwen3",
47
+ "num_attention_heads": 16,
48
+ "num_hidden_layers": 28,
49
+ "num_key_value_heads": 8,
50
+ "pad_token_id": 151643,
51
+ "rms_norm_eps": 1e-06,
52
+ "rope_scaling": null,
53
+ "rope_theta": 1000000,
54
+ "sliding_window": null,
55
+ "tie_word_embeddings": true,
56
+ "transformers_version": "4.57.3",
57
+ "use_cache": true,
58
+ "use_sliding_window": false,
59
+ "vocab_size": 151936
60
+ }
checkpoint-2300/generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token_id": [
3
+ 151643
4
+ ],
5
+ "max_new_tokens": 2048,
6
+ "pad_token_id": 151643,
7
+ "transformers_version": "4.57.3"
8
+ }
checkpoint-2300/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-2300/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2eb3c5c0658dbe15894c1fa31d764593a82789e005d7bb9bd6db1b4bb425b7b1
3
+ size 3441185608
checkpoint-2300/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00ba3deece7a14dcdf080ad9df36fbb8314b21b939405578a97ee97b1debded7
3
+ size 6882567855
checkpoint-2300/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fd9432b50864b1799d071a2391f13a188cc959e985675ab69fd688672db2853
3
+ size 14645
checkpoint-2300/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:782322f35441e7d1d0b7d391724a319d72286fd426f509707941177d09166877
3
+ size 1465
checkpoint-2300/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
checkpoint-2300/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
checkpoint-2300/tokenizer_config.json ADDED
@@ -0,0 +1,239 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|endoftext|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 131072,
235
+ "pad_token": "<|endoftext|>",
236
+ "split_special_tokens": false,
237
+ "tokenizer_class": "Qwen2Tokenizer",
238
+ "unk_token": null
239
+ }
checkpoint-2300/trainer_state.json ADDED
@@ -0,0 +1,2334 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 5.927835051546392,
6
+ "eval_steps": 500,
7
+ "global_step": 2300,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "entropy": 4.914482545852661,
14
+ "epoch": 0.02577319587628866,
15
+ "grad_norm": 232.0,
16
+ "learning_rate": 1.9313304721030046e-07,
17
+ "loss": 5.5997,
18
+ "mean_token_accuracy": 0.20726535096764565,
19
+ "num_tokens": 266.0,
20
+ "step": 10
21
+ },
22
+ {
23
+ "entropy": 4.538765811920166,
24
+ "epoch": 0.05154639175257732,
25
+ "grad_norm": 160.0,
26
+ "learning_rate": 4.07725321888412e-07,
27
+ "loss": 5.4934,
28
+ "mean_token_accuracy": 0.21887856498360633,
29
+ "num_tokens": 627.0,
30
+ "step": 20
31
+ },
32
+ {
33
+ "entropy": 4.939351224899292,
34
+ "epoch": 0.07731958762886598,
35
+ "grad_norm": 118.5,
36
+ "learning_rate": 6.223175965665236e-07,
37
+ "loss": 5.7523,
38
+ "mean_token_accuracy": 0.1917542487382889,
39
+ "num_tokens": 926.0,
40
+ "step": 30
41
+ },
42
+ {
43
+ "entropy": 4.698552560806275,
44
+ "epoch": 0.10309278350515463,
45
+ "grad_norm": 187.0,
46
+ "learning_rate": 8.369098712446352e-07,
47
+ "loss": 5.1204,
48
+ "mean_token_accuracy": 0.22331946194171906,
49
+ "num_tokens": 1259.0,
50
+ "step": 40
51
+ },
52
+ {
53
+ "entropy": 4.72123908996582,
54
+ "epoch": 0.12886597938144329,
55
+ "grad_norm": 166.0,
56
+ "learning_rate": 1.051502145922747e-06,
57
+ "loss": 5.198,
58
+ "mean_token_accuracy": 0.21775908395648003,
59
+ "num_tokens": 1550.0,
60
+ "step": 50
61
+ },
62
+ {
63
+ "entropy": 4.719539976119995,
64
+ "epoch": 0.15463917525773196,
65
+ "grad_norm": 200.0,
66
+ "learning_rate": 1.2660944206008586e-06,
67
+ "loss": 5.0832,
68
+ "mean_token_accuracy": 0.2405412092804909,
69
+ "num_tokens": 1839.0,
70
+ "step": 60
71
+ },
72
+ {
73
+ "entropy": 4.78384747505188,
74
+ "epoch": 0.18041237113402062,
75
+ "grad_norm": 149.0,
76
+ "learning_rate": 1.48068669527897e-06,
77
+ "loss": 5.0961,
78
+ "mean_token_accuracy": 0.19893446192145348,
79
+ "num_tokens": 2171.0,
80
+ "step": 70
81
+ },
82
+ {
83
+ "entropy": 4.974851608276367,
84
+ "epoch": 0.20618556701030927,
85
+ "grad_norm": 177.0,
86
+ "learning_rate": 1.6952789699570817e-06,
87
+ "loss": 5.2262,
88
+ "mean_token_accuracy": 0.20617640018463135,
89
+ "num_tokens": 2450.0,
90
+ "step": 80
91
+ },
92
+ {
93
+ "entropy": 4.705893039703369,
94
+ "epoch": 0.23195876288659795,
95
+ "grad_norm": 182.0,
96
+ "learning_rate": 1.9098712446351934e-06,
97
+ "loss": 4.8345,
98
+ "mean_token_accuracy": 0.22286981157958508,
99
+ "num_tokens": 2786.0,
100
+ "step": 90
101
+ },
102
+ {
103
+ "entropy": 4.696354675292969,
104
+ "epoch": 0.25773195876288657,
105
+ "grad_norm": 154.0,
106
+ "learning_rate": 2.124463519313305e-06,
107
+ "loss": 5.023,
108
+ "mean_token_accuracy": 0.25020611882209776,
109
+ "num_tokens": 3090.0,
110
+ "step": 100
111
+ },
112
+ {
113
+ "entropy": 4.626954746246338,
114
+ "epoch": 0.28350515463917525,
115
+ "grad_norm": 59.0,
116
+ "learning_rate": 2.3390557939914167e-06,
117
+ "loss": 4.5576,
118
+ "mean_token_accuracy": 0.2621785670518875,
119
+ "num_tokens": 3383.0,
120
+ "step": 110
121
+ },
122
+ {
123
+ "entropy": 4.682432174682617,
124
+ "epoch": 0.30927835051546393,
125
+ "grad_norm": 76.0,
126
+ "learning_rate": 2.553648068669528e-06,
127
+ "loss": 4.792,
128
+ "mean_token_accuracy": 0.26012246310710907,
129
+ "num_tokens": 3650.0,
130
+ "step": 120
131
+ },
132
+ {
133
+ "entropy": 4.563408708572387,
134
+ "epoch": 0.33505154639175255,
135
+ "grad_norm": 43.25,
136
+ "learning_rate": 2.7682403433476396e-06,
137
+ "loss": 4.5349,
138
+ "mean_token_accuracy": 0.22586977183818818,
139
+ "num_tokens": 3990.0,
140
+ "step": 130
141
+ },
142
+ {
143
+ "entropy": 4.270893955230713,
144
+ "epoch": 0.36082474226804123,
145
+ "grad_norm": 65.0,
146
+ "learning_rate": 2.982832618025751e-06,
147
+ "loss": 4.147,
148
+ "mean_token_accuracy": 0.28786555826663973,
149
+ "num_tokens": 4345.0,
150
+ "step": 140
151
+ },
152
+ {
153
+ "entropy": 4.0949385404586796,
154
+ "epoch": 0.3865979381443299,
155
+ "grad_norm": 55.75,
156
+ "learning_rate": 3.197424892703863e-06,
157
+ "loss": 3.7088,
158
+ "mean_token_accuracy": 0.34470676481723783,
159
+ "num_tokens": 4601.0,
160
+ "step": 150
161
+ },
162
+ {
163
+ "entropy": 4.128937864303589,
164
+ "epoch": 0.41237113402061853,
165
+ "grad_norm": 49.25,
166
+ "learning_rate": 3.412017167381975e-06,
167
+ "loss": 3.9797,
168
+ "mean_token_accuracy": 0.3258319616317749,
169
+ "num_tokens": 5024.0,
170
+ "step": 160
171
+ },
172
+ {
173
+ "entropy": 3.8451520442962646,
174
+ "epoch": 0.4381443298969072,
175
+ "grad_norm": 72.5,
176
+ "learning_rate": 3.6266094420600863e-06,
177
+ "loss": 3.7548,
178
+ "mean_token_accuracy": 0.3747887283563614,
179
+ "num_tokens": 5311.0,
180
+ "step": 170
181
+ },
182
+ {
183
+ "entropy": 3.712318539619446,
184
+ "epoch": 0.4639175257731959,
185
+ "grad_norm": 44.0,
186
+ "learning_rate": 3.841201716738197e-06,
187
+ "loss": 3.6765,
188
+ "mean_token_accuracy": 0.3758635461330414,
189
+ "num_tokens": 5651.0,
190
+ "step": 180
191
+ },
192
+ {
193
+ "entropy": 3.763609743118286,
194
+ "epoch": 0.4896907216494845,
195
+ "grad_norm": 43.25,
196
+ "learning_rate": 4.055793991416309e-06,
197
+ "loss": 3.6278,
198
+ "mean_token_accuracy": 0.3873393088579178,
199
+ "num_tokens": 5983.0,
200
+ "step": 190
201
+ },
202
+ {
203
+ "entropy": 3.781333661079407,
204
+ "epoch": 0.5154639175257731,
205
+ "grad_norm": 67.0,
206
+ "learning_rate": 4.270386266094421e-06,
207
+ "loss": 3.7018,
208
+ "mean_token_accuracy": 0.3747037798166275,
209
+ "num_tokens": 6296.0,
210
+ "step": 200
211
+ },
212
+ {
213
+ "entropy": 3.6542357921600344,
214
+ "epoch": 0.5412371134020618,
215
+ "grad_norm": 63.75,
216
+ "learning_rate": 4.484978540772533e-06,
217
+ "loss": 3.2866,
218
+ "mean_token_accuracy": 0.4405659481883049,
219
+ "num_tokens": 6658.0,
220
+ "step": 210
221
+ },
222
+ {
223
+ "entropy": 3.38469135761261,
224
+ "epoch": 0.5670103092783505,
225
+ "grad_norm": 88.0,
226
+ "learning_rate": 4.699570815450644e-06,
227
+ "loss": 3.1957,
228
+ "mean_token_accuracy": 0.41826934069395066,
229
+ "num_tokens": 6997.0,
230
+ "step": 220
231
+ },
232
+ {
233
+ "entropy": 3.9692420244216917,
234
+ "epoch": 0.5927835051546392,
235
+ "grad_norm": 42.5,
236
+ "learning_rate": 4.914163090128756e-06,
237
+ "loss": 4.0617,
238
+ "mean_token_accuracy": 0.3779368013143539,
239
+ "num_tokens": 7376.0,
240
+ "step": 230
241
+ },
242
+ {
243
+ "entropy": 3.725129175186157,
244
+ "epoch": 0.6185567010309279,
245
+ "grad_norm": 49.5,
246
+ "learning_rate": 4.999898809142829e-06,
247
+ "loss": 3.3293,
248
+ "mean_token_accuracy": 0.4062195152044296,
249
+ "num_tokens": 7688.0,
250
+ "step": 240
251
+ },
252
+ {
253
+ "entropy": 3.673329973220825,
254
+ "epoch": 0.6443298969072165,
255
+ "grad_norm": 84.5,
256
+ "learning_rate": 4.9992804502362914e-06,
257
+ "loss": 3.5039,
258
+ "mean_token_accuracy": 0.41747846007347106,
259
+ "num_tokens": 7988.0,
260
+ "step": 250
261
+ },
262
+ {
263
+ "entropy": 3.5831470012664797,
264
+ "epoch": 0.6701030927835051,
265
+ "grad_norm": 79.0,
266
+ "learning_rate": 4.998100088445351e-06,
267
+ "loss": 3.3134,
268
+ "mean_token_accuracy": 0.4633125364780426,
269
+ "num_tokens": 8263.0,
270
+ "step": 260
271
+ },
272
+ {
273
+ "entropy": 3.5552636861801146,
274
+ "epoch": 0.6958762886597938,
275
+ "grad_norm": 65.0,
276
+ "learning_rate": 4.996357989193094e-06,
277
+ "loss": 3.4462,
278
+ "mean_token_accuracy": 0.40612466633319855,
279
+ "num_tokens": 8590.0,
280
+ "step": 270
281
+ },
282
+ {
283
+ "entropy": 3.833337092399597,
284
+ "epoch": 0.7216494845360825,
285
+ "grad_norm": 39.5,
286
+ "learning_rate": 4.994054544218193e-06,
287
+ "loss": 3.7615,
288
+ "mean_token_accuracy": 0.3914962366223335,
289
+ "num_tokens": 8945.0,
290
+ "step": 280
291
+ },
292
+ {
293
+ "entropy": 3.6764284133911134,
294
+ "epoch": 0.7474226804123711,
295
+ "grad_norm": 60.25,
296
+ "learning_rate": 4.991190271486816e-06,
297
+ "loss": 3.4472,
298
+ "mean_token_accuracy": 0.4174343138933182,
299
+ "num_tokens": 9269.0,
300
+ "step": 290
301
+ },
302
+ {
303
+ "entropy": 3.5497223377227782,
304
+ "epoch": 0.7731958762886598,
305
+ "grad_norm": 76.0,
306
+ "learning_rate": 4.987765815076157e-06,
307
+ "loss": 3.1154,
308
+ "mean_token_accuracy": 0.44447133839130404,
309
+ "num_tokens": 9550.0,
310
+ "step": 300
311
+ },
312
+ {
313
+ "entropy": 3.8354516506195067,
314
+ "epoch": 0.7989690721649485,
315
+ "grad_norm": 58.25,
316
+ "learning_rate": 4.9837819450296e-06,
317
+ "loss": 3.4305,
318
+ "mean_token_accuracy": 0.42582335472106936,
319
+ "num_tokens": 9899.0,
320
+ "step": 310
321
+ },
322
+ {
323
+ "entropy": 3.752107357978821,
324
+ "epoch": 0.8247422680412371,
325
+ "grad_norm": 56.5,
326
+ "learning_rate": 4.979239557183571e-06,
327
+ "loss": 3.6763,
328
+ "mean_token_accuracy": 0.4198161542415619,
329
+ "num_tokens": 10174.0,
330
+ "step": 320
331
+ },
332
+ {
333
+ "entropy": 3.641463875770569,
334
+ "epoch": 0.8505154639175257,
335
+ "grad_norm": 61.25,
336
+ "learning_rate": 4.974139672966082e-06,
337
+ "loss": 3.5237,
338
+ "mean_token_accuracy": 0.3649302959442139,
339
+ "num_tokens": 10520.0,
340
+ "step": 330
341
+ },
342
+ {
343
+ "entropy": 3.487735724449158,
344
+ "epoch": 0.8762886597938144,
345
+ "grad_norm": 55.0,
346
+ "learning_rate": 4.968483439167061e-06,
347
+ "loss": 3.3832,
348
+ "mean_token_accuracy": 0.4202912449836731,
349
+ "num_tokens": 10834.0,
350
+ "step": 340
351
+ },
352
+ {
353
+ "entropy": 3.659078526496887,
354
+ "epoch": 0.9020618556701031,
355
+ "grad_norm": 58.0,
356
+ "learning_rate": 4.9622721276804674e-06,
357
+ "loss": 3.0981,
358
+ "mean_token_accuracy": 0.45194968581199646,
359
+ "num_tokens": 11110.0,
360
+ "step": 350
361
+ },
362
+ {
363
+ "entropy": 3.3279967069625855,
364
+ "epoch": 0.9278350515463918,
365
+ "grad_norm": 81.5,
366
+ "learning_rate": 4.955507135218291e-06,
367
+ "loss": 3.2825,
368
+ "mean_token_accuracy": 0.42670700550079343,
369
+ "num_tokens": 11410.0,
370
+ "step": 360
371
+ },
372
+ {
373
+ "entropy": 3.2990634202957154,
374
+ "epoch": 0.9536082474226805,
375
+ "grad_norm": 52.25,
376
+ "learning_rate": 4.948189982996479e-06,
377
+ "loss": 2.945,
378
+ "mean_token_accuracy": 0.46619434356689454,
379
+ "num_tokens": 11758.0,
380
+ "step": 370
381
+ },
382
+ {
383
+ "entropy": 3.442564105987549,
384
+ "epoch": 0.979381443298969,
385
+ "grad_norm": 77.0,
386
+ "learning_rate": 4.940322316392865e-06,
387
+ "loss": 3.4907,
388
+ "mean_token_accuracy": 0.465116411447525,
389
+ "num_tokens": 12008.0,
390
+ "step": 380
391
+ },
392
+ {
393
+ "entropy": 3.4102420568466187,
394
+ "epoch": 1.0051546391752577,
395
+ "grad_norm": 105.5,
396
+ "learning_rate": 4.931905904577182e-06,
397
+ "loss": 3.4858,
398
+ "mean_token_accuracy": 0.4123460859060287,
399
+ "num_tokens": 12288.0,
400
+ "step": 390
401
+ },
402
+ {
403
+ "entropy": 3.615120697021484,
404
+ "epoch": 1.0309278350515463,
405
+ "grad_norm": 57.75,
406
+ "learning_rate": 4.922942640113234e-06,
407
+ "loss": 3.1624,
408
+ "mean_token_accuracy": 0.45850674211978915,
409
+ "num_tokens": 12600.0,
410
+ "step": 400
411
+ },
412
+ {
413
+ "entropy": 3.5458458185195925,
414
+ "epoch": 1.056701030927835,
415
+ "grad_norm": 64.5,
416
+ "learning_rate": 4.913434538533324e-06,
417
+ "loss": 3.3334,
418
+ "mean_token_accuracy": 0.46345340013504027,
419
+ "num_tokens": 12888.0,
420
+ "step": 410
421
+ },
422
+ {
423
+ "entropy": 3.283875823020935,
424
+ "epoch": 1.0824742268041236,
425
+ "grad_norm": 84.5,
426
+ "learning_rate": 4.90338373788503e-06,
427
+ "loss": 3.2497,
428
+ "mean_token_accuracy": 0.47709590196609497,
429
+ "num_tokens": 13180.0,
430
+ "step": 420
431
+ },
432
+ {
433
+ "entropy": 3.4606318950653074,
434
+ "epoch": 1.1082474226804124,
435
+ "grad_norm": 72.0,
436
+ "learning_rate": 4.892792498250431e-06,
437
+ "loss": 3.5329,
438
+ "mean_token_accuracy": 0.4233353078365326,
439
+ "num_tokens": 13485.0,
440
+ "step": 430
441
+ },
442
+ {
443
+ "entropy": 3.3683727979660034,
444
+ "epoch": 1.134020618556701,
445
+ "grad_norm": 58.25,
446
+ "learning_rate": 4.881663201237889e-06,
447
+ "loss": 3.0689,
448
+ "mean_token_accuracy": 0.4563048958778381,
449
+ "num_tokens": 13849.0,
450
+ "step": 440
451
+ },
452
+ {
453
+ "entropy": 3.3520001411437987,
454
+ "epoch": 1.1597938144329896,
455
+ "grad_norm": 49.5,
456
+ "learning_rate": 4.869998349446514e-06,
457
+ "loss": 3.0789,
458
+ "mean_token_accuracy": 0.4614581674337387,
459
+ "num_tokens": 14121.0,
460
+ "step": 450
461
+ },
462
+ {
463
+ "entropy": 3.3299176692962646,
464
+ "epoch": 1.1855670103092784,
465
+ "grad_norm": 64.5,
466
+ "learning_rate": 4.857800565903405e-06,
467
+ "loss": 3.2253,
468
+ "mean_token_accuracy": 0.4312807470560074,
469
+ "num_tokens": 14503.0,
470
+ "step": 460
471
+ },
472
+ {
473
+ "entropy": 3.6635308504104613,
474
+ "epoch": 1.211340206185567,
475
+ "grad_norm": 52.5,
476
+ "learning_rate": 4.845072593473826e-06,
477
+ "loss": 3.2584,
478
+ "mean_token_accuracy": 0.46550854444503786,
479
+ "num_tokens": 14792.0,
480
+ "step": 470
481
+ },
482
+ {
483
+ "entropy": 3.2862602949142454,
484
+ "epoch": 1.2371134020618557,
485
+ "grad_norm": 56.25,
486
+ "learning_rate": 4.831817294244432e-06,
487
+ "loss": 3.1376,
488
+ "mean_token_accuracy": 0.4864483565092087,
489
+ "num_tokens": 15071.0,
490
+ "step": 480
491
+ },
492
+ {
493
+ "entropy": 3.4670738697052004,
494
+ "epoch": 1.2628865979381443,
495
+ "grad_norm": 60.25,
496
+ "learning_rate": 4.8180376488796755e-06,
497
+ "loss": 3.4598,
498
+ "mean_token_accuracy": 0.38327425718307495,
499
+ "num_tokens": 15340.0,
500
+ "step": 490
501
+ },
502
+ {
503
+ "entropy": 3.4762473344802856,
504
+ "epoch": 1.2886597938144329,
505
+ "grad_norm": 72.0,
506
+ "learning_rate": 4.803736755951564e-06,
507
+ "loss": 3.23,
508
+ "mean_token_accuracy": 0.4640804290771484,
509
+ "num_tokens": 15658.0,
510
+ "step": 500
511
+ },
512
+ {
513
+ "entropy": 3.4152861833572388,
514
+ "epoch": 1.3144329896907216,
515
+ "grad_norm": 60.5,
516
+ "learning_rate": 4.788917831242895e-06,
517
+ "loss": 3.1239,
518
+ "mean_token_accuracy": 0.47079411447048186,
519
+ "num_tokens": 15943.0,
520
+ "step": 510
521
+ },
522
+ {
523
+ "entropy": 3.3919612884521486,
524
+ "epoch": 1.3402061855670104,
525
+ "grad_norm": 116.5,
526
+ "learning_rate": 4.773584207024135e-06,
527
+ "loss": 3.193,
528
+ "mean_token_accuracy": 0.40636845529079435,
529
+ "num_tokens": 16268.0,
530
+ "step": 520
531
+ },
532
+ {
533
+ "entropy": 3.520834803581238,
534
+ "epoch": 1.365979381443299,
535
+ "grad_norm": 67.5,
536
+ "learning_rate": 4.7577393313041025e-06,
537
+ "loss": 3.3889,
538
+ "mean_token_accuracy": 0.4428008824586868,
539
+ "num_tokens": 16611.0,
540
+ "step": 530
541
+ },
542
+ {
543
+ "entropy": 3.5736474275588987,
544
+ "epoch": 1.3917525773195876,
545
+ "grad_norm": 51.25,
546
+ "learning_rate": 4.741386767054636e-06,
547
+ "loss": 3.4062,
548
+ "mean_token_accuracy": 0.408548378944397,
549
+ "num_tokens": 16917.0,
550
+ "step": 540
551
+ },
552
+ {
553
+ "entropy": 3.4252942323684694,
554
+ "epoch": 1.4175257731958764,
555
+ "grad_norm": 66.5,
556
+ "learning_rate": 4.724530191409399e-06,
557
+ "loss": 3.2558,
558
+ "mean_token_accuracy": 0.4369684547185898,
559
+ "num_tokens": 17283.0,
560
+ "step": 550
561
+ },
562
+ {
563
+ "entropy": 3.335077738761902,
564
+ "epoch": 1.443298969072165,
565
+ "grad_norm": 57.25,
566
+ "learning_rate": 4.707173394837017e-06,
567
+ "loss": 3.157,
568
+ "mean_token_accuracy": 0.4764990329742432,
569
+ "num_tokens": 17604.0,
570
+ "step": 560
571
+ },
572
+ {
573
+ "entropy": 3.519501209259033,
574
+ "epoch": 1.4690721649484537,
575
+ "grad_norm": 57.5,
576
+ "learning_rate": 4.689320280288731e-06,
577
+ "loss": 3.4428,
578
+ "mean_token_accuracy": 0.4320195406675339,
579
+ "num_tokens": 17939.0,
580
+ "step": 570
581
+ },
582
+ {
583
+ "entropy": 3.5304367780685424,
584
+ "epoch": 1.4948453608247423,
585
+ "grad_norm": 55.0,
586
+ "learning_rate": 4.67097486232076e-06,
587
+ "loss": 3.2563,
588
+ "mean_token_accuracy": 0.4069958388805389,
589
+ "num_tokens": 18234.0,
590
+ "step": 580
591
+ },
592
+ {
593
+ "entropy": 3.662277579307556,
594
+ "epoch": 1.5206185567010309,
595
+ "grad_norm": 67.0,
596
+ "learning_rate": 4.65214126619156e-06,
597
+ "loss": 3.8052,
598
+ "mean_token_accuracy": 0.40810766220092776,
599
+ "num_tokens": 18574.0,
600
+ "step": 590
601
+ },
602
+ {
603
+ "entropy": 3.2554187536239625,
604
+ "epoch": 1.5463917525773194,
605
+ "grad_norm": 69.0,
606
+ "learning_rate": 4.632823726934199e-06,
607
+ "loss": 3.0088,
608
+ "mean_token_accuracy": 0.4515380173921585,
609
+ "num_tokens": 18886.0,
610
+ "step": 600
611
+ },
612
+ {
613
+ "entropy": 3.4018918752670286,
614
+ "epoch": 1.5721649484536082,
615
+ "grad_norm": 60.25,
616
+ "learning_rate": 4.613026588404036e-06,
617
+ "loss": 3.3767,
618
+ "mean_token_accuracy": 0.4256245791912079,
619
+ "num_tokens": 19227.0,
620
+ "step": 610
621
+ },
622
+ {
623
+ "entropy": 3.4922634601593017,
624
+ "epoch": 1.597938144329897,
625
+ "grad_norm": 45.5,
626
+ "learning_rate": 4.592754302301942e-06,
627
+ "loss": 3.0169,
628
+ "mean_token_accuracy": 0.4801509857177734,
629
+ "num_tokens": 19582.0,
630
+ "step": 620
631
+ },
632
+ {
633
+ "entropy": 3.3631627798080443,
634
+ "epoch": 1.6237113402061856,
635
+ "grad_norm": 74.0,
636
+ "learning_rate": 4.572011427173263e-06,
637
+ "loss": 3.2208,
638
+ "mean_token_accuracy": 0.4472234547138214,
639
+ "num_tokens": 19934.0,
640
+ "step": 630
641
+ },
642
+ {
643
+ "entropy": 3.353942036628723,
644
+ "epoch": 1.6494845360824741,
645
+ "grad_norm": 54.25,
646
+ "learning_rate": 4.550802627382756e-06,
647
+ "loss": 3.2015,
648
+ "mean_token_accuracy": 0.4314853399991989,
649
+ "num_tokens": 20286.0,
650
+ "step": 640
651
+ },
652
+ {
653
+ "entropy": 3.338911604881287,
654
+ "epoch": 1.675257731958763,
655
+ "grad_norm": 89.0,
656
+ "learning_rate": 4.529132672065738e-06,
657
+ "loss": 3.09,
658
+ "mean_token_accuracy": 0.46670873165130616,
659
+ "num_tokens": 20556.0,
660
+ "step": 650
661
+ },
662
+ {
663
+ "entropy": 3.380303430557251,
664
+ "epoch": 1.7010309278350515,
665
+ "grad_norm": 76.0,
666
+ "learning_rate": 4.507006434055663e-06,
667
+ "loss": 3.2405,
668
+ "mean_token_accuracy": 0.4376333147287369,
669
+ "num_tokens": 20862.0,
670
+ "step": 660
671
+ },
672
+ {
673
+ "entropy": 3.491442632675171,
674
+ "epoch": 1.7268041237113403,
675
+ "grad_norm": 50.25,
676
+ "learning_rate": 4.484428888788395e-06,
677
+ "loss": 3.4112,
678
+ "mean_token_accuracy": 0.43711408972740173,
679
+ "num_tokens": 21164.0,
680
+ "step": 670
681
+ },
682
+ {
683
+ "entropy": 3.451434350013733,
684
+ "epoch": 1.7525773195876289,
685
+ "grad_norm": 105.5,
686
+ "learning_rate": 4.461405113183396e-06,
687
+ "loss": 3.0601,
688
+ "mean_token_accuracy": 0.42957675755023955,
689
+ "num_tokens": 21460.0,
690
+ "step": 680
691
+ },
692
+ {
693
+ "entropy": 3.329232668876648,
694
+ "epoch": 1.7783505154639174,
695
+ "grad_norm": 63.25,
696
+ "learning_rate": 4.437940284502105e-06,
697
+ "loss": 3.1711,
698
+ "mean_token_accuracy": 0.4769585371017456,
699
+ "num_tokens": 21817.0,
700
+ "step": 690
701
+ },
702
+ {
703
+ "entropy": 3.1766634225845336,
704
+ "epoch": 1.8041237113402062,
705
+ "grad_norm": 47.25,
706
+ "learning_rate": 4.414039679183749e-06,
707
+ "loss": 3.5047,
708
+ "mean_token_accuracy": 0.43050127625465395,
709
+ "num_tokens": 22163.0,
710
+ "step": 700
711
+ },
712
+ {
713
+ "entropy": 3.394679617881775,
714
+ "epoch": 1.829896907216495,
715
+ "grad_norm": 64.0,
716
+ "learning_rate": 4.389708671658844e-06,
717
+ "loss": 3.0371,
718
+ "mean_token_accuracy": 0.45515852570533755,
719
+ "num_tokens": 22436.0,
720
+ "step": 710
721
+ },
722
+ {
723
+ "entropy": 3.4196753025054933,
724
+ "epoch": 1.8556701030927836,
725
+ "grad_norm": 217.0,
726
+ "learning_rate": 4.3649527331406796e-06,
727
+ "loss": 3.612,
728
+ "mean_token_accuracy": 0.4677813768386841,
729
+ "num_tokens": 22709.0,
730
+ "step": 720
731
+ },
732
+ {
733
+ "entropy": 3.4829108476638795,
734
+ "epoch": 1.8814432989690721,
735
+ "grad_norm": 60.25,
736
+ "learning_rate": 4.339777430395022e-06,
737
+ "loss": 3.3854,
738
+ "mean_token_accuracy": 0.41964206099510193,
739
+ "num_tokens": 23076.0,
740
+ "step": 730
741
+ },
742
+ {
743
+ "entropy": 3.3097528219223022,
744
+ "epoch": 1.9072164948453607,
745
+ "grad_norm": 63.25,
746
+ "learning_rate": 4.314188424488344e-06,
747
+ "loss": 2.8523,
748
+ "mean_token_accuracy": 0.48268236219882965,
749
+ "num_tokens": 23403.0,
750
+ "step": 740
751
+ },
752
+ {
753
+ "entropy": 3.3857040882110594,
754
+ "epoch": 1.9329896907216495,
755
+ "grad_norm": 65.5,
756
+ "learning_rate": 4.288191469514839e-06,
757
+ "loss": 2.9944,
758
+ "mean_token_accuracy": 0.43578424155712125,
759
+ "num_tokens": 23722.0,
760
+ "step": 750
761
+ },
762
+ {
763
+ "entropy": 3.408646750450134,
764
+ "epoch": 1.9587628865979383,
765
+ "grad_norm": 68.5,
766
+ "learning_rate": 4.261792411302525e-06,
767
+ "loss": 3.4775,
768
+ "mean_token_accuracy": 0.43519311845302583,
769
+ "num_tokens": 23993.0,
770
+ "step": 760
771
+ },
772
+ {
773
+ "entropy": 3.117679166793823,
774
+ "epoch": 1.9845360824742269,
775
+ "grad_norm": 57.25,
776
+ "learning_rate": 4.234997186098716e-06,
777
+ "loss": 2.829,
778
+ "mean_token_accuracy": 0.43563964366912844,
779
+ "num_tokens": 24324.0,
780
+ "step": 770
781
+ },
782
+ {
783
+ "entropy": 3.5846817255020142,
784
+ "epoch": 2.0103092783505154,
785
+ "grad_norm": 41.5,
786
+ "learning_rate": 4.207811819235164e-06,
787
+ "loss": 3.374,
788
+ "mean_token_accuracy": 0.4384703665971756,
789
+ "num_tokens": 24690.0,
790
+ "step": 780
791
+ },
792
+ {
793
+ "entropy": 3.0135667085647584,
794
+ "epoch": 2.036082474226804,
795
+ "grad_norm": 49.25,
796
+ "learning_rate": 4.180242423773166e-06,
797
+ "loss": 2.9591,
798
+ "mean_token_accuracy": 0.5096089750528335,
799
+ "num_tokens": 24959.0,
800
+ "step": 790
801
+ },
802
+ {
803
+ "entropy": 3.289955735206604,
804
+ "epoch": 2.0618556701030926,
805
+ "grad_norm": 52.75,
806
+ "learning_rate": 4.1522951991289465e-06,
807
+ "loss": 3.218,
808
+ "mean_token_accuracy": 0.39164299368858335,
809
+ "num_tokens": 25285.0,
810
+ "step": 800
811
+ },
812
+ {
813
+ "entropy": 3.3328727006912233,
814
+ "epoch": 2.0876288659793816,
815
+ "grad_norm": 58.25,
816
+ "learning_rate": 4.1239764296796175e-06,
817
+ "loss": 2.9698,
818
+ "mean_token_accuracy": 0.4740764260292053,
819
+ "num_tokens": 25578.0,
820
+ "step": 810
821
+ },
822
+ {
823
+ "entropy": 3.3463461637496947,
824
+ "epoch": 2.11340206185567,
825
+ "grad_norm": 74.0,
826
+ "learning_rate": 4.095292483350041e-06,
827
+ "loss": 3.2668,
828
+ "mean_token_accuracy": 0.41175087094306945,
829
+ "num_tokens": 25954.0,
830
+ "step": 820
831
+ },
832
+ {
833
+ "entropy": 3.347863268852234,
834
+ "epoch": 2.1391752577319587,
835
+ "grad_norm": 93.5,
836
+ "learning_rate": 4.066249810180895e-06,
837
+ "loss": 3.1272,
838
+ "mean_token_accuracy": 0.44678041338920593,
839
+ "num_tokens": 26225.0,
840
+ "step": 830
841
+ },
842
+ {
843
+ "entropy": 3.2035139322280886,
844
+ "epoch": 2.1649484536082473,
845
+ "grad_norm": 69.0,
846
+ "learning_rate": 4.036854940878284e-06,
847
+ "loss": 3.2895,
848
+ "mean_token_accuracy": 0.41500436663627627,
849
+ "num_tokens": 26615.0,
850
+ "step": 840
851
+ },
852
+ {
853
+ "entropy": 3.2597701787948608,
854
+ "epoch": 2.1907216494845363,
855
+ "grad_norm": 49.0,
856
+ "learning_rate": 4.007114485345205e-06,
857
+ "loss": 3.0027,
858
+ "mean_token_accuracy": 0.529634228348732,
859
+ "num_tokens": 26948.0,
860
+ "step": 850
861
+ },
862
+ {
863
+ "entropy": 3.3507386445999146,
864
+ "epoch": 2.216494845360825,
865
+ "grad_norm": 54.0,
866
+ "learning_rate": 3.977035131195202e-06,
867
+ "loss": 3.0962,
868
+ "mean_token_accuracy": 0.4740526854991913,
869
+ "num_tokens": 27268.0,
870
+ "step": 860
871
+ },
872
+ {
873
+ "entropy": 3.2031197786331176,
874
+ "epoch": 2.2422680412371134,
875
+ "grad_norm": 81.5,
876
+ "learning_rate": 3.946623642248554e-06,
877
+ "loss": 2.5841,
878
+ "mean_token_accuracy": 0.5312987476587295,
879
+ "num_tokens": 27549.0,
880
+ "step": 870
881
+ },
882
+ {
883
+ "entropy": 3.2550023555755616,
884
+ "epoch": 2.268041237113402,
885
+ "grad_norm": 69.5,
886
+ "learning_rate": 3.915886857011323e-06,
887
+ "loss": 3.1698,
888
+ "mean_token_accuracy": 0.4532750606536865,
889
+ "num_tokens": 27886.0,
890
+ "step": 880
891
+ },
892
+ {
893
+ "entropy": 3.485032653808594,
894
+ "epoch": 2.2938144329896906,
895
+ "grad_norm": 55.5,
896
+ "learning_rate": 3.8848316871376055e-06,
897
+ "loss": 3.4646,
898
+ "mean_token_accuracy": 0.4280081331729889,
899
+ "num_tokens": 28200.0,
900
+ "step": 890
901
+ },
902
+ {
903
+ "entropy": 3.3659554958343505,
904
+ "epoch": 2.319587628865979,
905
+ "grad_norm": 97.5,
906
+ "learning_rate": 3.853465115875335e-06,
907
+ "loss": 3.2682,
908
+ "mean_token_accuracy": 0.45062295794487,
909
+ "num_tokens": 28536.0,
910
+ "step": 900
911
+ },
912
+ {
913
+ "entropy": 3.13352746963501,
914
+ "epoch": 2.345360824742268,
915
+ "grad_norm": 80.0,
916
+ "learning_rate": 3.821794196495995e-06,
917
+ "loss": 2.9256,
918
+ "mean_token_accuracy": 0.4756322205066681,
919
+ "num_tokens": 28810.0,
920
+ "step": 910
921
+ },
922
+ {
923
+ "entropy": 3.135547161102295,
924
+ "epoch": 2.3711340206185567,
925
+ "grad_norm": 65.5,
926
+ "learning_rate": 3.7898260507085697e-06,
927
+ "loss": 2.9811,
928
+ "mean_token_accuracy": 0.4623017519712448,
929
+ "num_tokens": 29089.0,
930
+ "step": 920
931
+ },
932
+ {
933
+ "entropy": 3.449264574050903,
934
+ "epoch": 2.3969072164948453,
935
+ "grad_norm": 61.75,
936
+ "learning_rate": 3.757567867058125e-06,
937
+ "loss": 3.6536,
938
+ "mean_token_accuracy": 0.40380783975124357,
939
+ "num_tokens": 29403.0,
940
+ "step": 930
941
+ },
942
+ {
943
+ "entropy": 3.4234744548797607,
944
+ "epoch": 2.422680412371134,
945
+ "grad_norm": 61.5,
946
+ "learning_rate": 3.7250268993093396e-06,
947
+ "loss": 3.2193,
948
+ "mean_token_accuracy": 0.49216817915439603,
949
+ "num_tokens": 29689.0,
950
+ "step": 940
951
+ },
952
+ {
953
+ "entropy": 3.4770421504974367,
954
+ "epoch": 2.448453608247423,
955
+ "grad_norm": 44.0,
956
+ "learning_rate": 3.6922104648153885e-06,
957
+ "loss": 3.3501,
958
+ "mean_token_accuracy": 0.4592595547437668,
959
+ "num_tokens": 30047.0,
960
+ "step": 950
961
+ },
962
+ {
963
+ "entropy": 3.339312434196472,
964
+ "epoch": 2.4742268041237114,
965
+ "grad_norm": 55.0,
966
+ "learning_rate": 3.659125942872516e-06,
967
+ "loss": 3.075,
968
+ "mean_token_accuracy": 0.4719044387340546,
969
+ "num_tokens": 30376.0,
970
+ "step": 960
971
+ },
972
+ {
973
+ "entropy": 3.4688915252685546,
974
+ "epoch": 2.5,
975
+ "grad_norm": 61.75,
976
+ "learning_rate": 3.625780773060687e-06,
977
+ "loss": 3.1916,
978
+ "mean_token_accuracy": 0.41194990575313567,
979
+ "num_tokens": 30691.0,
980
+ "step": 970
981
+ },
982
+ {
983
+ "entropy": 3.302336239814758,
984
+ "epoch": 2.5257731958762886,
985
+ "grad_norm": 67.5,
986
+ "learning_rate": 3.5921824535706756e-06,
987
+ "loss": 3.1441,
988
+ "mean_token_accuracy": 0.4681535869836807,
989
+ "num_tokens": 31007.0,
990
+ "step": 980
991
+ },
992
+ {
993
+ "entropy": 3.343183135986328,
994
+ "epoch": 2.551546391752577,
995
+ "grad_norm": 71.0,
996
+ "learning_rate": 3.5583385395179793e-06,
997
+ "loss": 3.1333,
998
+ "mean_token_accuracy": 0.47766990661621095,
999
+ "num_tokens": 31272.0,
1000
+ "step": 990
1001
+ },
1002
+ {
1003
+ "entropy": 3.13364782333374,
1004
+ "epoch": 2.5773195876288657,
1005
+ "grad_norm": 80.5,
1006
+ "learning_rate": 3.5242566412439332e-06,
1007
+ "loss": 3.1096,
1008
+ "mean_token_accuracy": 0.46111657917499543,
1009
+ "num_tokens": 31549.0,
1010
+ "step": 1000
1011
+ },
1012
+ {
1013
+ "entropy": 3.2046863794326783,
1014
+ "epoch": 2.6030927835051547,
1015
+ "grad_norm": 103.0,
1016
+ "learning_rate": 3.4899444226044023e-06,
1017
+ "loss": 3.5947,
1018
+ "mean_token_accuracy": 0.41809163987636566,
1019
+ "num_tokens": 31839.0,
1020
+ "step": 1010
1021
+ },
1022
+ {
1023
+ "entropy": 3.2269564390182497,
1024
+ "epoch": 2.6288659793814433,
1025
+ "grad_norm": 60.0,
1026
+ "learning_rate": 3.455409599246442e-06,
1027
+ "loss": 3.0418,
1028
+ "mean_token_accuracy": 0.4730542838573456,
1029
+ "num_tokens": 32131.0,
1030
+ "step": 1020
1031
+ },
1032
+ {
1033
+ "entropy": 3.255638074874878,
1034
+ "epoch": 2.654639175257732,
1035
+ "grad_norm": 53.25,
1036
+ "learning_rate": 3.4206599368733114e-06,
1037
+ "loss": 3.1486,
1038
+ "mean_token_accuracy": 0.48221423029899596,
1039
+ "num_tokens": 32472.0,
1040
+ "step": 1030
1041
+ },
1042
+ {
1043
+ "entropy": 3.304290223121643,
1044
+ "epoch": 2.680412371134021,
1045
+ "grad_norm": 51.5,
1046
+ "learning_rate": 3.3857032494982327e-06,
1047
+ "loss": 3.1768,
1048
+ "mean_token_accuracy": 0.44427731931209563,
1049
+ "num_tokens": 32882.0,
1050
+ "step": 1040
1051
+ },
1052
+ {
1053
+ "entropy": 3.4136642932891847,
1054
+ "epoch": 2.7061855670103094,
1055
+ "grad_norm": 53.25,
1056
+ "learning_rate": 3.3505473976872883e-06,
1057
+ "loss": 3.1119,
1058
+ "mean_token_accuracy": 0.45050349533557893,
1059
+ "num_tokens": 33165.0,
1060
+ "step": 1050
1061
+ },
1062
+ {
1063
+ "entropy": 3.386446309089661,
1064
+ "epoch": 2.731958762886598,
1065
+ "grad_norm": 71.0,
1066
+ "learning_rate": 3.3152002867918433e-06,
1067
+ "loss": 3.2298,
1068
+ "mean_token_accuracy": 0.41270871758461,
1069
+ "num_tokens": 33441.0,
1070
+ "step": 1060
1071
+ },
1072
+ {
1073
+ "entropy": 3.571580958366394,
1074
+ "epoch": 2.7577319587628866,
1075
+ "grad_norm": 59.5,
1076
+ "learning_rate": 3.279669865170906e-06,
1077
+ "loss": 3.3263,
1078
+ "mean_token_accuracy": 0.45758517682552335,
1079
+ "num_tokens": 33761.0,
1080
+ "step": 1070
1081
+ },
1082
+ {
1083
+ "entropy": 3.489269471168518,
1084
+ "epoch": 2.783505154639175,
1085
+ "grad_norm": 78.0,
1086
+ "learning_rate": 3.2439641224038093e-06,
1087
+ "loss": 3.0612,
1088
+ "mean_token_accuracy": 0.4716973781585693,
1089
+ "num_tokens": 34086.0,
1090
+ "step": 1080
1091
+ },
1092
+ {
1093
+ "entropy": 3.352162575721741,
1094
+ "epoch": 2.8092783505154637,
1095
+ "grad_norm": 56.0,
1096
+ "learning_rate": 3.20809108749363e-06,
1097
+ "loss": 3.1394,
1098
+ "mean_token_accuracy": 0.45618820786476133,
1099
+ "num_tokens": 34422.0,
1100
+ "step": 1090
1101
+ },
1102
+ {
1103
+ "entropy": 3.4569438695907593,
1104
+ "epoch": 2.8350515463917527,
1105
+ "grad_norm": 76.5,
1106
+ "learning_rate": 3.17205882706174e-06,
1107
+ "loss": 3.0988,
1108
+ "mean_token_accuracy": 0.48512459397315977,
1109
+ "num_tokens": 34726.0,
1110
+ "step": 1100
1111
+ },
1112
+ {
1113
+ "entropy": 3.376886796951294,
1114
+ "epoch": 2.8608247422680413,
1115
+ "grad_norm": 103.0,
1116
+ "learning_rate": 3.135875443533896e-06,
1117
+ "loss": 3.4308,
1118
+ "mean_token_accuracy": 0.41293925344944,
1119
+ "num_tokens": 35057.0,
1120
+ "step": 1110
1121
+ },
1122
+ {
1123
+ "entropy": 3.207944130897522,
1124
+ "epoch": 2.88659793814433,
1125
+ "grad_norm": 72.5,
1126
+ "learning_rate": 3.0995490733182825e-06,
1127
+ "loss": 2.7498,
1128
+ "mean_token_accuracy": 0.47106904685497286,
1129
+ "num_tokens": 35327.0,
1130
+ "step": 1120
1131
+ },
1132
+ {
1133
+ "entropy": 3.2526353359222413,
1134
+ "epoch": 2.9123711340206184,
1135
+ "grad_norm": 73.5,
1136
+ "learning_rate": 3.06308788497591e-06,
1137
+ "loss": 3.1798,
1138
+ "mean_token_accuracy": 0.4563720256090164,
1139
+ "num_tokens": 35617.0,
1140
+ "step": 1130
1141
+ },
1142
+ {
1143
+ "entropy": 3.3556177854537963,
1144
+ "epoch": 2.9381443298969074,
1145
+ "grad_norm": 67.0,
1146
+ "learning_rate": 3.026500077383785e-06,
1147
+ "loss": 3.1254,
1148
+ "mean_token_accuracy": 0.45203186869621276,
1149
+ "num_tokens": 35941.0,
1150
+ "step": 1140
1151
+ },
1152
+ {
1153
+ "entropy": 3.4933328866958617,
1154
+ "epoch": 2.963917525773196,
1155
+ "grad_norm": 65.5,
1156
+ "learning_rate": 2.989793877891263e-06,
1157
+ "loss": 3.5153,
1158
+ "mean_token_accuracy": 0.38476662933826444,
1159
+ "num_tokens": 36320.0,
1160
+ "step": 1150
1161
+ },
1162
+ {
1163
+ "entropy": 3.307176351547241,
1164
+ "epoch": 2.9896907216494846,
1165
+ "grad_norm": 80.5,
1166
+ "learning_rate": 2.952977540469999e-06,
1167
+ "loss": 3.0624,
1168
+ "mean_token_accuracy": 0.4267964720726013,
1169
+ "num_tokens": 36595.0,
1170
+ "step": 1160
1171
+ },
1172
+ {
1173
+ "entropy": 3.25522928237915,
1174
+ "epoch": 3.015463917525773,
1175
+ "grad_norm": 71.5,
1176
+ "learning_rate": 2.9160593438579054e-06,
1177
+ "loss": 2.8878,
1178
+ "mean_token_accuracy": 0.4728631258010864,
1179
+ "num_tokens": 36963.0,
1180
+ "step": 1170
1181
+ },
1182
+ {
1183
+ "entropy": 3.160351610183716,
1184
+ "epoch": 3.0412371134020617,
1185
+ "grad_norm": 73.5,
1186
+ "learning_rate": 2.879047589697555e-06,
1187
+ "loss": 3.0562,
1188
+ "mean_token_accuracy": 0.5218929082155228,
1189
+ "num_tokens": 37236.0,
1190
+ "step": 1180
1191
+ },
1192
+ {
1193
+ "entropy": 3.497001218795776,
1194
+ "epoch": 3.0670103092783507,
1195
+ "grad_norm": 91.0,
1196
+ "learning_rate": 2.8419506006694186e-06,
1197
+ "loss": 3.2947,
1198
+ "mean_token_accuracy": 0.40737067759037016,
1199
+ "num_tokens": 37522.0,
1200
+ "step": 1190
1201
+ },
1202
+ {
1203
+ "entropy": 3.3410105228424074,
1204
+ "epoch": 3.0927835051546393,
1205
+ "grad_norm": 61.5,
1206
+ "learning_rate": 2.8047767186203808e-06,
1207
+ "loss": 3.2214,
1208
+ "mean_token_accuracy": 0.45971156358718873,
1209
+ "num_tokens": 37825.0,
1210
+ "step": 1200
1211
+ },
1212
+ {
1213
+ "entropy": 3.4304397821426393,
1214
+ "epoch": 3.118556701030928,
1215
+ "grad_norm": 68.0,
1216
+ "learning_rate": 2.767534302687942e-06,
1217
+ "loss": 3.5544,
1218
+ "mean_token_accuracy": 0.42803180813789365,
1219
+ "num_tokens": 38181.0,
1220
+ "step": 1210
1221
+ },
1222
+ {
1223
+ "entropy": 3.2708693027496336,
1224
+ "epoch": 3.1443298969072164,
1225
+ "grad_norm": 51.5,
1226
+ "learning_rate": 2.730231727420533e-06,
1227
+ "loss": 3.0771,
1228
+ "mean_token_accuracy": 0.42246846556663514,
1229
+ "num_tokens": 38521.0,
1230
+ "step": 1220
1231
+ },
1232
+ {
1233
+ "entropy": 3.278656005859375,
1234
+ "epoch": 3.170103092783505,
1235
+ "grad_norm": 58.75,
1236
+ "learning_rate": 2.6928773808943696e-06,
1237
+ "loss": 3.2341,
1238
+ "mean_token_accuracy": 0.4356242328882217,
1239
+ "num_tokens": 38806.0,
1240
+ "step": 1230
1241
+ },
1242
+ {
1243
+ "entropy": 3.4282007455825805,
1244
+ "epoch": 3.195876288659794,
1245
+ "grad_norm": 85.5,
1246
+ "learning_rate": 2.6554796628272534e-06,
1247
+ "loss": 3.3791,
1248
+ "mean_token_accuracy": 0.44304960370063784,
1249
+ "num_tokens": 39104.0,
1250
+ "step": 1240
1251
+ },
1252
+ {
1253
+ "entropy": 3.3579065799713135,
1254
+ "epoch": 3.2216494845360826,
1255
+ "grad_norm": 68.5,
1256
+ "learning_rate": 2.6180469826897683e-06,
1257
+ "loss": 3.3142,
1258
+ "mean_token_accuracy": 0.41488229632377627,
1259
+ "num_tokens": 39433.0,
1260
+ "step": 1250
1261
+ },
1262
+ {
1263
+ "entropy": 3.2836158752441404,
1264
+ "epoch": 3.247422680412371,
1265
+ "grad_norm": 58.25,
1266
+ "learning_rate": 2.5805877578142713e-06,
1267
+ "loss": 3.0095,
1268
+ "mean_token_accuracy": 0.47964567244052886,
1269
+ "num_tokens": 39730.0,
1270
+ "step": 1260
1271
+ },
1272
+ {
1273
+ "entropy": 3.3673797607421876,
1274
+ "epoch": 3.2731958762886597,
1275
+ "grad_norm": 76.0,
1276
+ "learning_rate": 2.543110411502129e-06,
1277
+ "loss": 2.7924,
1278
+ "mean_token_accuracy": 0.46048580706119535,
1279
+ "num_tokens": 40029.0,
1280
+ "step": 1270
1281
+ },
1282
+ {
1283
+ "entropy": 3.2721627712249757,
1284
+ "epoch": 3.2989690721649483,
1285
+ "grad_norm": 64.5,
1286
+ "learning_rate": 2.5056233711295985e-06,
1287
+ "loss": 3.0903,
1288
+ "mean_token_accuracy": 0.4774274632334709,
1289
+ "num_tokens": 40310.0,
1290
+ "step": 1280
1291
+ },
1292
+ {
1293
+ "entropy": 3.1327761888504027,
1294
+ "epoch": 3.3247422680412373,
1295
+ "grad_norm": 78.0,
1296
+ "learning_rate": 2.4681350662528004e-06,
1297
+ "loss": 2.9106,
1298
+ "mean_token_accuracy": 0.45230706930160525,
1299
+ "num_tokens": 40609.0,
1300
+ "step": 1290
1301
+ },
1302
+ {
1303
+ "entropy": 3.2892409801483153,
1304
+ "epoch": 3.350515463917526,
1305
+ "grad_norm": 89.5,
1306
+ "learning_rate": 2.4306539267122e-06,
1307
+ "loss": 3.2275,
1308
+ "mean_token_accuracy": 0.42299139201641084,
1309
+ "num_tokens": 40899.0,
1310
+ "step": 1300
1311
+ },
1312
+ {
1313
+ "entropy": 3.4429001808166504,
1314
+ "epoch": 3.3762886597938144,
1315
+ "grad_norm": 55.5,
1316
+ "learning_rate": 2.393188380737021e-06,
1317
+ "loss": 3.1632,
1318
+ "mean_token_accuracy": 0.45537458956241605,
1319
+ "num_tokens": 41229.0,
1320
+ "step": 1310
1321
+ },
1322
+ {
1323
+ "entropy": 3.1667943239212035,
1324
+ "epoch": 3.402061855670103,
1325
+ "grad_norm": 61.25,
1326
+ "learning_rate": 2.3557468530500298e-06,
1327
+ "loss": 3.0767,
1328
+ "mean_token_accuracy": 0.41762115657329557,
1329
+ "num_tokens": 41595.0,
1330
+ "step": 1320
1331
+ },
1332
+ {
1333
+ "entropy": 2.933236765861511,
1334
+ "epoch": 3.4278350515463916,
1335
+ "grad_norm": 51.0,
1336
+ "learning_rate": 2.3183377629730963e-06,
1337
+ "loss": 2.5435,
1338
+ "mean_token_accuracy": 0.507614666223526,
1339
+ "num_tokens": 41919.0,
1340
+ "step": 1330
1341
+ },
1342
+ {
1343
+ "entropy": 3.458651614189148,
1344
+ "epoch": 3.4536082474226806,
1345
+ "grad_norm": 57.5,
1346
+ "learning_rate": 2.28096952253398e-06,
1347
+ "loss": 3.842,
1348
+ "mean_token_accuracy": 0.39763966798782346,
1349
+ "num_tokens": 42237.0,
1350
+ "step": 1340
1351
+ },
1352
+ {
1353
+ "entropy": 3.4523784637451174,
1354
+ "epoch": 3.479381443298969,
1355
+ "grad_norm": 43.5,
1356
+ "learning_rate": 2.2436505345747505e-06,
1357
+ "loss": 3.1589,
1358
+ "mean_token_accuracy": 0.4656664371490479,
1359
+ "num_tokens": 42575.0,
1360
+ "step": 1350
1361
+ },
1362
+ {
1363
+ "entropy": 3.239696431159973,
1364
+ "epoch": 3.5051546391752577,
1365
+ "grad_norm": 74.0,
1366
+ "learning_rate": 2.2063891908622767e-06,
1367
+ "loss": 2.7964,
1368
+ "mean_token_accuracy": 0.48300274908542634,
1369
+ "num_tokens": 42936.0,
1370
+ "step": 1360
1371
+ },
1372
+ {
1373
+ "entropy": 3.3155257225036623,
1374
+ "epoch": 3.5309278350515463,
1375
+ "grad_norm": 58.75,
1376
+ "learning_rate": 2.169193870201203e-06,
1377
+ "loss": 3.106,
1378
+ "mean_token_accuracy": 0.463299959897995,
1379
+ "num_tokens": 43256.0,
1380
+ "step": 1370
1381
+ },
1382
+ {
1383
+ "entropy": 3.298970675468445,
1384
+ "epoch": 3.556701030927835,
1385
+ "grad_norm": 99.5,
1386
+ "learning_rate": 2.1320729365498404e-06,
1387
+ "loss": 3.3008,
1388
+ "mean_token_accuracy": 0.4579169362783432,
1389
+ "num_tokens": 43519.0,
1390
+ "step": 1380
1391
+ },
1392
+ {
1393
+ "entropy": 3.373097324371338,
1394
+ "epoch": 3.582474226804124,
1395
+ "grad_norm": 61.0,
1396
+ "learning_rate": 2.095034737139404e-06,
1397
+ "loss": 3.068,
1398
+ "mean_token_accuracy": 0.4262126713991165,
1399
+ "num_tokens": 43814.0,
1400
+ "step": 1390
1401
+ },
1402
+ {
1403
+ "entropy": 3.1966411590576174,
1404
+ "epoch": 3.6082474226804124,
1405
+ "grad_norm": 70.5,
1406
+ "learning_rate": 2.058087600596997e-06,
1407
+ "loss": 2.8911,
1408
+ "mean_token_accuracy": 0.5069397330284119,
1409
+ "num_tokens": 44153.0,
1410
+ "step": 1400
1411
+ },
1412
+ {
1413
+ "entropy": 3.3376646757125856,
1414
+ "epoch": 3.634020618556701,
1415
+ "grad_norm": 64.5,
1416
+ "learning_rate": 2.021239835072794e-06,
1417
+ "loss": 3.0766,
1418
+ "mean_token_accuracy": 0.46726988852024076,
1419
+ "num_tokens": 44465.0,
1420
+ "step": 1410
1421
+ },
1422
+ {
1423
+ "entropy": 3.3123555183410645,
1424
+ "epoch": 3.6597938144329896,
1425
+ "grad_norm": 53.5,
1426
+ "learning_rate": 1.984499726371819e-06,
1427
+ "loss": 2.8527,
1428
+ "mean_token_accuracy": 0.5056732088327408,
1429
+ "num_tokens": 44792.0,
1430
+ "step": 1420
1431
+ },
1432
+ {
1433
+ "entropy": 3.2575713872909544,
1434
+ "epoch": 3.6855670103092786,
1435
+ "grad_norm": 92.0,
1436
+ "learning_rate": 1.947875536090748e-06,
1437
+ "loss": 3.0459,
1438
+ "mean_token_accuracy": 0.46606201231479644,
1439
+ "num_tokens": 45093.0,
1440
+ "step": 1430
1441
+ },
1442
+ {
1443
+ "entropy": 3.2526177883148195,
1444
+ "epoch": 3.711340206185567,
1445
+ "grad_norm": 56.0,
1446
+ "learning_rate": 1.9113754997601614e-06,
1447
+ "loss": 3.2644,
1448
+ "mean_token_accuracy": 0.4528957188129425,
1449
+ "num_tokens": 45381.0,
1450
+ "step": 1440
1451
+ },
1452
+ {
1453
+ "entropy": 3.166100788116455,
1454
+ "epoch": 3.7371134020618557,
1455
+ "grad_norm": 60.75,
1456
+ "learning_rate": 1.875007824992654e-06,
1457
+ "loss": 2.9971,
1458
+ "mean_token_accuracy": 0.4395914673805237,
1459
+ "num_tokens": 45723.0,
1460
+ "step": 1450
1461
+ },
1462
+ {
1463
+ "entropy": 3.41998028755188,
1464
+ "epoch": 3.7628865979381443,
1465
+ "grad_norm": 75.0,
1466
+ "learning_rate": 1.8387806896372206e-06,
1467
+ "loss": 3.2834,
1468
+ "mean_token_accuracy": 0.4282302588224411,
1469
+ "num_tokens": 46053.0,
1470
+ "step": 1460
1471
+ },
1472
+ {
1473
+ "entropy": 3.263835144042969,
1474
+ "epoch": 3.788659793814433,
1475
+ "grad_norm": 73.0,
1476
+ "learning_rate": 1.8027022399403377e-06,
1477
+ "loss": 3.1976,
1478
+ "mean_token_accuracy": 0.4730199307203293,
1479
+ "num_tokens": 46330.0,
1480
+ "step": 1470
1481
+ },
1482
+ {
1483
+ "entropy": 3.38959059715271,
1484
+ "epoch": 3.8144329896907214,
1485
+ "grad_norm": 86.5,
1486
+ "learning_rate": 1.7667805887141526e-06,
1487
+ "loss": 3.072,
1488
+ "mean_token_accuracy": 0.48471441566944123,
1489
+ "num_tokens": 46675.0,
1490
+ "step": 1480
1491
+ },
1492
+ {
1493
+ "entropy": 3.516207385063171,
1494
+ "epoch": 3.8402061855670104,
1495
+ "grad_norm": 69.5,
1496
+ "learning_rate": 1.731023813512186e-06,
1497
+ "loss": 3.1848,
1498
+ "mean_token_accuracy": 0.4615781009197235,
1499
+ "num_tokens": 47005.0,
1500
+ "step": 1490
1501
+ },
1502
+ {
1503
+ "entropy": 3.435403323173523,
1504
+ "epoch": 3.865979381443299,
1505
+ "grad_norm": 62.5,
1506
+ "learning_rate": 1.695439954812968e-06,
1507
+ "loss": 3.5288,
1508
+ "mean_token_accuracy": 0.4129458874464035,
1509
+ "num_tokens": 47320.0,
1510
+ "step": 1500
1511
+ },
1512
+ {
1513
+ "entropy": 3.0035399436950683,
1514
+ "epoch": 3.8917525773195876,
1515
+ "grad_norm": 87.5,
1516
+ "learning_rate": 1.660037014212009e-06,
1517
+ "loss": 2.8988,
1518
+ "mean_token_accuracy": 0.5202886313199997,
1519
+ "num_tokens": 47654.0,
1520
+ "step": 1510
1521
+ },
1522
+ {
1523
+ "entropy": 3.282400608062744,
1524
+ "epoch": 3.917525773195876,
1525
+ "grad_norm": 68.0,
1526
+ "learning_rate": 1.62482295262251e-06,
1527
+ "loss": 3.0969,
1528
+ "mean_token_accuracy": 0.42603813409805297,
1529
+ "num_tokens": 47988.0,
1530
+ "step": 1520
1531
+ },
1532
+ {
1533
+ "entropy": 3.5067455053329466,
1534
+ "epoch": 3.943298969072165,
1535
+ "grad_norm": 63.75,
1536
+ "learning_rate": 1.589805688485231e-06,
1537
+ "loss": 3.2709,
1538
+ "mean_token_accuracy": 0.4377409517765045,
1539
+ "num_tokens": 48303.0,
1540
+ "step": 1530
1541
+ },
1542
+ {
1543
+ "entropy": 3.245515561103821,
1544
+ "epoch": 3.9690721649484537,
1545
+ "grad_norm": 55.0,
1546
+ "learning_rate": 1.5549930959878996e-06,
1547
+ "loss": 2.9502,
1548
+ "mean_token_accuracy": 0.5143730461597442,
1549
+ "num_tokens": 48637.0,
1550
+ "step": 1540
1551
+ },
1552
+ {
1553
+ "entropy": 3.1619110345840453,
1554
+ "epoch": 3.9948453608247423,
1555
+ "grad_norm": 61.25,
1556
+ "learning_rate": 1.5203930032945765e-06,
1557
+ "loss": 2.883,
1558
+ "mean_token_accuracy": 0.4930521368980408,
1559
+ "num_tokens": 48955.0,
1560
+ "step": 1550
1561
+ },
1562
+ {
1563
+ "entropy": 3.368650484085083,
1564
+ "epoch": 4.020618556701031,
1565
+ "grad_norm": 59.75,
1566
+ "learning_rate": 1.4860131907853664e-06,
1567
+ "loss": 3.0012,
1568
+ "mean_token_accuracy": 0.4367584019899368,
1569
+ "num_tokens": 49264.0,
1570
+ "step": 1560
1571
+ },
1572
+ {
1573
+ "entropy": 3.2799517631530763,
1574
+ "epoch": 4.046391752577319,
1575
+ "grad_norm": 107.0,
1576
+ "learning_rate": 1.45186138930688e-06,
1577
+ "loss": 3.0185,
1578
+ "mean_token_accuracy": 0.45549334287643434,
1579
+ "num_tokens": 49528.0,
1580
+ "step": 1570
1581
+ },
1582
+ {
1583
+ "entropy": 3.293484115600586,
1584
+ "epoch": 4.072164948453608,
1585
+ "grad_norm": 70.0,
1586
+ "learning_rate": 1.4179452784338265e-06,
1587
+ "loss": 3.3586,
1588
+ "mean_token_accuracy": 0.41464213728904725,
1589
+ "num_tokens": 49823.0,
1590
+ "step": 1580
1591
+ },
1592
+ {
1593
+ "entropy": 3.212877941131592,
1594
+ "epoch": 4.097938144329897,
1595
+ "grad_norm": 58.0,
1596
+ "learning_rate": 1.3842724847421435e-06,
1597
+ "loss": 2.9933,
1598
+ "mean_token_accuracy": 0.41281671822071075,
1599
+ "num_tokens": 50209.0,
1600
+ "step": 1590
1601
+ },
1602
+ {
1603
+ "entropy": 3.2077365159988402,
1604
+ "epoch": 4.123711340206185,
1605
+ "grad_norm": 60.0,
1606
+ "learning_rate": 1.3508505800940327e-06,
1607
+ "loss": 3.2048,
1608
+ "mean_token_accuracy": 0.45703212916851044,
1609
+ "num_tokens": 50566.0,
1610
+ "step": 1600
1611
+ },
1612
+ {
1613
+ "entropy": 3.2609221935272217,
1614
+ "epoch": 4.149484536082475,
1615
+ "grad_norm": 87.5,
1616
+ "learning_rate": 1.317687079935317e-06,
1617
+ "loss": 3.2903,
1618
+ "mean_token_accuracy": 0.45516538321971894,
1619
+ "num_tokens": 50932.0,
1620
+ "step": 1610
1621
+ },
1622
+ {
1623
+ "entropy": 3.199517750740051,
1624
+ "epoch": 4.175257731958763,
1625
+ "grad_norm": 121.5,
1626
+ "learning_rate": 1.2847894416054645e-06,
1627
+ "loss": 2.9641,
1628
+ "mean_token_accuracy": 0.49091013371944425,
1629
+ "num_tokens": 51242.0,
1630
+ "step": 1620
1631
+ },
1632
+ {
1633
+ "entropy": 3.2298731803894043,
1634
+ "epoch": 4.201030927835052,
1635
+ "grad_norm": 62.75,
1636
+ "learning_rate": 1.2521650626606926e-06,
1637
+ "loss": 2.9035,
1638
+ "mean_token_accuracy": 0.5309109538793564,
1639
+ "num_tokens": 51519.0,
1640
+ "step": 1630
1641
+ },
1642
+ {
1643
+ "entropy": 3.3670015573501586,
1644
+ "epoch": 4.22680412371134,
1645
+ "grad_norm": 80.0,
1646
+ "learning_rate": 1.219821279210507e-06,
1647
+ "loss": 3.0805,
1648
+ "mean_token_accuracy": 0.444789519906044,
1649
+ "num_tokens": 51823.0,
1650
+ "step": 1640
1651
+ },
1652
+ {
1653
+ "entropy": 3.3619128465652466,
1654
+ "epoch": 4.252577319587629,
1655
+ "grad_norm": 89.5,
1656
+ "learning_rate": 1.1877653642680618e-06,
1657
+ "loss": 3.0084,
1658
+ "mean_token_accuracy": 0.45184328258037565,
1659
+ "num_tokens": 52106.0,
1660
+ "step": 1650
1661
+ },
1662
+ {
1663
+ "entropy": 3.4395506143569947,
1664
+ "epoch": 4.278350515463917,
1665
+ "grad_norm": 72.0,
1666
+ "learning_rate": 1.1560045261147079e-06,
1667
+ "loss": 3.4761,
1668
+ "mean_token_accuracy": 0.41549868881702423,
1669
+ "num_tokens": 52494.0,
1670
+ "step": 1660
1671
+ },
1672
+ {
1673
+ "entropy": 3.154675102233887,
1674
+ "epoch": 4.304123711340206,
1675
+ "grad_norm": 62.25,
1676
+ "learning_rate": 1.1245459066790962e-06,
1677
+ "loss": 2.9656,
1678
+ "mean_token_accuracy": 0.49724196195602416,
1679
+ "num_tokens": 52842.0,
1680
+ "step": 1670
1681
+ },
1682
+ {
1683
+ "entropy": 3.107476019859314,
1684
+ "epoch": 4.329896907216495,
1685
+ "grad_norm": 87.5,
1686
+ "learning_rate": 1.0933965799312015e-06,
1687
+ "loss": 2.8218,
1688
+ "mean_token_accuracy": 0.5045855909585952,
1689
+ "num_tokens": 53174.0,
1690
+ "step": 1680
1691
+ },
1692
+ {
1693
+ "entropy": 3.290882110595703,
1694
+ "epoch": 4.355670103092783,
1695
+ "grad_norm": 87.0,
1696
+ "learning_rate": 1.062563550291626e-06,
1697
+ "loss": 3.0095,
1698
+ "mean_token_accuracy": 0.48295737206935885,
1699
+ "num_tokens": 53410.0,
1700
+ "step": 1690
1701
+ },
1702
+ {
1703
+ "entropy": 3.23433403968811,
1704
+ "epoch": 4.381443298969073,
1705
+ "grad_norm": 57.5,
1706
+ "learning_rate": 1.0320537510565474e-06,
1707
+ "loss": 3.0136,
1708
+ "mean_token_accuracy": 0.4336048990488052,
1709
+ "num_tokens": 53745.0,
1710
+ "step": 1700
1711
+ },
1712
+ {
1713
+ "entropy": 3.3636296510696413,
1714
+ "epoch": 4.407216494845361,
1715
+ "grad_norm": 60.5,
1716
+ "learning_rate": 1.0018740428386562e-06,
1717
+ "loss": 3.0679,
1718
+ "mean_token_accuracy": 0.47437537312507627,
1719
+ "num_tokens": 54031.0,
1720
+ "step": 1710
1721
+ },
1722
+ {
1723
+ "entropy": 3.443432426452637,
1724
+ "epoch": 4.43298969072165,
1725
+ "grad_norm": 86.5,
1726
+ "learning_rate": 9.720312120244368e-07,
1727
+ "loss": 3.2067,
1728
+ "mean_token_accuracy": 0.4474021762609482,
1729
+ "num_tokens": 54356.0,
1730
+ "step": 1720
1731
+ },
1732
+ {
1733
+ "entropy": 3.4575706720352173,
1734
+ "epoch": 4.458762886597938,
1735
+ "grad_norm": 80.0,
1736
+ "learning_rate": 9.425319692481421e-07,
1737
+ "loss": 3.244,
1738
+ "mean_token_accuracy": 0.44102594554424285,
1739
+ "num_tokens": 54627.0,
1740
+ "step": 1730
1741
+ },
1742
+ {
1743
+ "entropy": 3.0440264463424684,
1744
+ "epoch": 4.484536082474227,
1745
+ "grad_norm": 63.0,
1746
+ "learning_rate": 9.133829478828e-07,
1747
+ "loss": 2.9248,
1748
+ "mean_token_accuracy": 0.5230745673179626,
1749
+ "num_tokens": 54941.0,
1750
+ "step": 1740
1751
+ },
1752
+ {
1753
+ "entropy": 3.0223827600479125,
1754
+ "epoch": 4.510309278350515,
1755
+ "grad_norm": 77.5,
1756
+ "learning_rate": 8.845907025485945e-07,
1757
+ "loss": 2.8445,
1758
+ "mean_token_accuracy": 0.5213218927383423,
1759
+ "num_tokens": 55227.0,
1760
+ "step": 1750
1761
+ },
1762
+ {
1763
+ "entropy": 3.4792301177978517,
1764
+ "epoch": 4.536082474226804,
1765
+ "grad_norm": 63.0,
1766
+ "learning_rate": 8.561617076389556e-07,
1767
+ "loss": 3.1867,
1768
+ "mean_token_accuracy": 0.4527750164270401,
1769
+ "num_tokens": 55547.0,
1770
+ "step": 1760
1771
+ },
1772
+ {
1773
+ "entropy": 3.3560336589813233,
1774
+ "epoch": 4.561855670103093,
1775
+ "grad_norm": 85.0,
1776
+ "learning_rate": 8.281023558646892e-07,
1777
+ "loss": 3.2218,
1778
+ "mean_token_accuracy": 0.4219012975692749,
1779
+ "num_tokens": 55880.0,
1780
+ "step": 1770
1781
+ },
1782
+ {
1783
+ "entropy": 3.2443851470947265,
1784
+ "epoch": 4.587628865979381,
1785
+ "grad_norm": 72.0,
1786
+ "learning_rate": 8.004189568164721e-07,
1787
+ "loss": 3.2558,
1788
+ "mean_token_accuracy": 0.4505341827869415,
1789
+ "num_tokens": 56225.0,
1790
+ "step": 1780
1791
+ },
1792
+ {
1793
+ "entropy": 3.254337120056152,
1794
+ "epoch": 4.61340206185567,
1795
+ "grad_norm": 75.5,
1796
+ "learning_rate": 7.731177355460456e-07,
1797
+ "loss": 3.1863,
1798
+ "mean_token_accuracy": 0.4375568628311157,
1799
+ "num_tokens": 56528.0,
1800
+ "step": 1790
1801
+ },
1802
+ {
1803
+ "entropy": 3.0034953951835632,
1804
+ "epoch": 4.639175257731958,
1805
+ "grad_norm": 64.0,
1806
+ "learning_rate": 7.462048311664086e-07,
1807
+ "loss": 2.6958,
1808
+ "mean_token_accuracy": 0.536104878783226,
1809
+ "num_tokens": 56847.0,
1810
+ "step": 1800
1811
+ },
1812
+ {
1813
+ "entropy": 3.2068035364151,
1814
+ "epoch": 4.664948453608248,
1815
+ "grad_norm": 85.0,
1816
+ "learning_rate": 7.196862954713438e-07,
1817
+ "loss": 2.9987,
1818
+ "mean_token_accuracy": 0.4882556527853012,
1819
+ "num_tokens": 57154.0,
1820
+ "step": 1810
1821
+ },
1822
+ {
1823
+ "entropy": 3.0603734254837036,
1824
+ "epoch": 4.690721649484536,
1825
+ "grad_norm": 54.5,
1826
+ "learning_rate": 6.935680915745743e-07,
1827
+ "loss": 2.6614,
1828
+ "mean_token_accuracy": 0.4894832164049149,
1829
+ "num_tokens": 57445.0,
1830
+ "step": 1820
1831
+ },
1832
+ {
1833
+ "entropy": 3.4436551094055177,
1834
+ "epoch": 4.716494845360825,
1835
+ "grad_norm": 66.0,
1836
+ "learning_rate": 6.678560925688629e-07,
1837
+ "loss": 3.4491,
1838
+ "mean_token_accuracy": 0.42854584753513336,
1839
+ "num_tokens": 57799.0,
1840
+ "step": 1830
1841
+ },
1842
+ {
1843
+ "entropy": 3.3284261226654053,
1844
+ "epoch": 4.742268041237113,
1845
+ "grad_norm": 69.5,
1846
+ "learning_rate": 6.425560802053551e-07,
1847
+ "loss": 3.1212,
1848
+ "mean_token_accuracy": 0.4246213287115097,
1849
+ "num_tokens": 58153.0,
1850
+ "step": 1840
1851
+ },
1852
+ {
1853
+ "entropy": 3.3664702415466308,
1854
+ "epoch": 4.768041237113402,
1855
+ "grad_norm": 85.5,
1856
+ "learning_rate": 6.176737435934593e-07,
1857
+ "loss": 2.97,
1858
+ "mean_token_accuracy": 0.45907922089099884,
1859
+ "num_tokens": 58473.0,
1860
+ "step": 1850
1861
+ },
1862
+ {
1863
+ "entropy": 3.260459637641907,
1864
+ "epoch": 4.793814432989691,
1865
+ "grad_norm": 70.0,
1866
+ "learning_rate": 5.932146779215614e-07,
1867
+ "loss": 3.091,
1868
+ "mean_token_accuracy": 0.47388018369674684,
1869
+ "num_tokens": 58787.0,
1870
+ "step": 1860
1871
+ },
1872
+ {
1873
+ "entropy": 3.260399317741394,
1874
+ "epoch": 4.819587628865979,
1875
+ "grad_norm": 64.0,
1876
+ "learning_rate": 5.691843831988547e-07,
1877
+ "loss": 3.1014,
1878
+ "mean_token_accuracy": 0.4497444421052933,
1879
+ "num_tokens": 59126.0,
1880
+ "step": 1870
1881
+ },
1882
+ {
1883
+ "entropy": 3.1505658864974975,
1884
+ "epoch": 4.845360824742268,
1885
+ "grad_norm": 101.5,
1886
+ "learning_rate": 5.45588263018581e-07,
1887
+ "loss": 3.0726,
1888
+ "mean_token_accuracy": 0.4976365238428116,
1889
+ "num_tokens": 59480.0,
1890
+ "step": 1880
1891
+ },
1892
+ {
1893
+ "entropy": 3.3082144498825072,
1894
+ "epoch": 4.871134020618557,
1895
+ "grad_norm": 55.25,
1896
+ "learning_rate": 5.224316233429422e-07,
1897
+ "loss": 3.4489,
1898
+ "mean_token_accuracy": 0.41699815094470977,
1899
+ "num_tokens": 59755.0,
1900
+ "step": 1890
1901
+ },
1902
+ {
1903
+ "entropy": 3.4492565393447876,
1904
+ "epoch": 4.896907216494846,
1905
+ "grad_norm": 129.0,
1906
+ "learning_rate": 4.997196713099728e-07,
1907
+ "loss": 3.3794,
1908
+ "mean_token_accuracy": 0.44576013684272764,
1909
+ "num_tokens": 60046.0,
1910
+ "step": 1900
1911
+ },
1912
+ {
1913
+ "entropy": 3.398744559288025,
1914
+ "epoch": 4.922680412371134,
1915
+ "grad_norm": 88.0,
1916
+ "learning_rate": 4.774575140626317e-07,
1917
+ "loss": 3.3877,
1918
+ "mean_token_accuracy": 0.4413867056369781,
1919
+ "num_tokens": 60384.0,
1920
+ "step": 1910
1921
+ },
1922
+ {
1923
+ "entropy": 3.20350604057312,
1924
+ "epoch": 4.948453608247423,
1925
+ "grad_norm": 56.0,
1926
+ "learning_rate": 4.556501576003791e-07,
1927
+ "loss": 3.0359,
1928
+ "mean_token_accuracy": 0.45435314774513247,
1929
+ "num_tokens": 60655.0,
1930
+ "step": 1920
1931
+ },
1932
+ {
1933
+ "entropy": 2.980875849723816,
1934
+ "epoch": 4.974226804123711,
1935
+ "grad_norm": 80.5,
1936
+ "learning_rate": 4.343025056534994e-07,
1937
+ "loss": 2.9326,
1938
+ "mean_token_accuracy": 0.5152811527252197,
1939
+ "num_tokens": 60949.0,
1940
+ "step": 1930
1941
+ },
1942
+ {
1943
+ "entropy": 3.3593399286270142,
1944
+ "epoch": 5.0,
1945
+ "grad_norm": 70.5,
1946
+ "learning_rate": 4.134193585804197e-07,
1947
+ "loss": 3.2147,
1948
+ "mean_token_accuracy": 0.4529120117425919,
1949
+ "num_tokens": 61260.0,
1950
+ "step": 1940
1951
+ },
1952
+ {
1953
+ "entropy": 3.3000378370285035,
1954
+ "epoch": 5.025773195876289,
1955
+ "grad_norm": 75.0,
1956
+ "learning_rate": 3.930054122882709e-07,
1957
+ "loss": 3.2406,
1958
+ "mean_token_accuracy": 0.4164623826742172,
1959
+ "num_tokens": 61558.0,
1960
+ "step": 1950
1961
+ },
1962
+ {
1963
+ "entropy": 3.211987781524658,
1964
+ "epoch": 5.051546391752577,
1965
+ "grad_norm": 56.75,
1966
+ "learning_rate": 3.7306525717694195e-07,
1967
+ "loss": 3.1386,
1968
+ "mean_token_accuracy": 0.5111187756061554,
1969
+ "num_tokens": 61848.0,
1970
+ "step": 1960
1971
+ },
1972
+ {
1973
+ "entropy": 3.2421340942382812,
1974
+ "epoch": 5.077319587628866,
1975
+ "grad_norm": 79.5,
1976
+ "learning_rate": 3.536033771068506e-07,
1977
+ "loss": 2.9573,
1978
+ "mean_token_accuracy": 0.49586873054504393,
1979
+ "num_tokens": 62126.0,
1980
+ "step": 1970
1981
+ },
1982
+ {
1983
+ "entropy": 3.501763868331909,
1984
+ "epoch": 5.103092783505154,
1985
+ "grad_norm": 67.5,
1986
+ "learning_rate": 3.3462414839068233e-07,
1987
+ "loss": 3.2128,
1988
+ "mean_token_accuracy": 0.4240656703710556,
1989
+ "num_tokens": 62529.0,
1990
+ "step": 1980
1991
+ },
1992
+ {
1993
+ "entropy": 3.2703943729400633,
1994
+ "epoch": 5.128865979381443,
1995
+ "grad_norm": 57.5,
1996
+ "learning_rate": 3.1613183880930124e-07,
1997
+ "loss": 3.1511,
1998
+ "mean_token_accuracy": 0.4509860217571259,
1999
+ "num_tokens": 62808.0,
2000
+ "step": 1990
2001
+ },
2002
+ {
2003
+ "entropy": 3.1918478488922117,
2004
+ "epoch": 5.154639175257732,
2005
+ "grad_norm": 90.0,
2006
+ "learning_rate": 2.9813060665207615e-07,
2007
+ "loss": 2.9956,
2008
+ "mean_token_accuracy": 0.5052088230848313,
2009
+ "num_tokens": 63054.0,
2010
+ "step": 2000
2011
+ },
2012
+ {
2013
+ "entropy": 3.2867510318756104,
2014
+ "epoch": 5.180412371134021,
2015
+ "grad_norm": 62.0,
2016
+ "learning_rate": 2.8062449978182056e-07,
2017
+ "loss": 3.2784,
2018
+ "mean_token_accuracy": 0.4433626294136047,
2019
+ "num_tokens": 63378.0,
2020
+ "step": 2010
2021
+ },
2022
+ {
2023
+ "entropy": 3.276996612548828,
2024
+ "epoch": 5.206185567010309,
2025
+ "grad_norm": 59.5,
2026
+ "learning_rate": 2.636174547245671e-07,
2027
+ "loss": 2.9721,
2028
+ "mean_token_accuracy": 0.4633091241121292,
2029
+ "num_tokens": 63698.0,
2030
+ "step": 2020
2031
+ },
2032
+ {
2033
+ "entropy": 3.3924339771270753,
2034
+ "epoch": 5.231958762886598,
2035
+ "grad_norm": 70.5,
2036
+ "learning_rate": 2.471132957843775e-07,
2037
+ "loss": 3.5247,
2038
+ "mean_token_accuracy": 0.40621828436851504,
2039
+ "num_tokens": 64026.0,
2040
+ "step": 2030
2041
+ },
2042
+ {
2043
+ "entropy": 3.183038115501404,
2044
+ "epoch": 5.257731958762887,
2045
+ "grad_norm": 64.0,
2046
+ "learning_rate": 2.3111573418338724e-07,
2047
+ "loss": 3.057,
2048
+ "mean_token_accuracy": 0.4755241394042969,
2049
+ "num_tokens": 64357.0,
2050
+ "step": 2040
2051
+ },
2052
+ {
2053
+ "entropy": 3.0321537017822267,
2054
+ "epoch": 5.283505154639175,
2055
+ "grad_norm": 80.0,
2056
+ "learning_rate": 2.156283672272777e-07,
2057
+ "loss": 3.0536,
2058
+ "mean_token_accuracy": 0.46814982295036317,
2059
+ "num_tokens": 64717.0,
2060
+ "step": 2050
2061
+ },
2062
+ {
2063
+ "entropy": 3.307017612457275,
2064
+ "epoch": 5.309278350515464,
2065
+ "grad_norm": 50.5,
2066
+ "learning_rate": 2.0065467749636497e-07,
2067
+ "loss": 3.0379,
2068
+ "mean_token_accuracy": 0.4797358334064484,
2069
+ "num_tokens": 65004.0,
2070
+ "step": 2060
2071
+ },
2072
+ {
2073
+ "entropy": 3.342577505111694,
2074
+ "epoch": 5.335051546391752,
2075
+ "grad_norm": 103.5,
2076
+ "learning_rate": 1.861980320624873e-07,
2077
+ "loss": 3.111,
2078
+ "mean_token_accuracy": 0.46001616716384885,
2079
+ "num_tokens": 65327.0,
2080
+ "step": 2070
2081
+ },
2082
+ {
2083
+ "entropy": 3.2767983198165895,
2084
+ "epoch": 5.360824742268041,
2085
+ "grad_norm": 51.25,
2086
+ "learning_rate": 1.7226168173186396e-07,
2087
+ "loss": 3.2626,
2088
+ "mean_token_accuracy": 0.44220549464225767,
2089
+ "num_tokens": 65743.0,
2090
+ "step": 2080
2091
+ },
2092
+ {
2093
+ "entropy": 3.1632169485092163,
2094
+ "epoch": 5.38659793814433,
2095
+ "grad_norm": 80.0,
2096
+ "learning_rate": 1.5884876031410202e-07,
2097
+ "loss": 3.4269,
2098
+ "mean_token_accuracy": 0.43744454979896547,
2099
+ "num_tokens": 66042.0,
2100
+ "step": 2090
2101
+ },
2102
+ {
2103
+ "entropy": 3.083737111091614,
2104
+ "epoch": 5.412371134020619,
2105
+ "grad_norm": 76.0,
2106
+ "learning_rate": 1.4596228391750676e-07,
2107
+ "loss": 3.0883,
2108
+ "mean_token_accuracy": 0.4829647660255432,
2109
+ "num_tokens": 66327.0,
2110
+ "step": 2100
2111
+ },
2112
+ {
2113
+ "entropy": 3.032327151298523,
2114
+ "epoch": 5.438144329896907,
2115
+ "grad_norm": 99.5,
2116
+ "learning_rate": 1.3360515027086462e-07,
2117
+ "loss": 2.8775,
2118
+ "mean_token_accuracy": 0.4383985221385956,
2119
+ "num_tokens": 66641.0,
2120
+ "step": 2110
2121
+ },
2122
+ {
2123
+ "entropy": 3.05249879360199,
2124
+ "epoch": 5.463917525773196,
2125
+ "grad_norm": 87.0,
2126
+ "learning_rate": 1.2178013807184085e-07,
2127
+ "loss": 3.0836,
2128
+ "mean_token_accuracy": 0.457595694065094,
2129
+ "num_tokens": 66937.0,
2130
+ "step": 2120
2131
+ },
2132
+ {
2133
+ "entropy": 3.3152157068252563,
2134
+ "epoch": 5.489690721649485,
2135
+ "grad_norm": 76.0,
2136
+ "learning_rate": 1.1048990636214618e-07,
2137
+ "loss": 3.1917,
2138
+ "mean_token_accuracy": 0.48176924884319305,
2139
+ "num_tokens": 67223.0,
2140
+ "step": 2130
2141
+ },
2142
+ {
2143
+ "entropy": 3.5079283475875855,
2144
+ "epoch": 5.515463917525773,
2145
+ "grad_norm": 60.75,
2146
+ "learning_rate": 9.973699392960917e-08,
2147
+ "loss": 3.3596,
2148
+ "mean_token_accuracy": 0.40318597555160524,
2149
+ "num_tokens": 67536.0,
2150
+ "step": 2140
2151
+ },
2152
+ {
2153
+ "entropy": 3.342143940925598,
2154
+ "epoch": 5.541237113402062,
2155
+ "grad_norm": 65.5,
2156
+ "learning_rate": 8.952381873728821e-08,
2157
+ "loss": 2.9796,
2158
+ "mean_token_accuracy": 0.46548409163951876,
2159
+ "num_tokens": 67876.0,
2160
+ "step": 2150
2161
+ },
2162
+ {
2163
+ "entropy": 3.1259532451629637,
2164
+ "epoch": 5.56701030927835,
2165
+ "grad_norm": 49.75,
2166
+ "learning_rate": 7.985267737975588e-08,
2167
+ "loss": 3.0859,
2168
+ "mean_token_accuracy": 0.4911260426044464,
2169
+ "num_tokens": 68229.0,
2170
+ "step": 2160
2171
+ },
2172
+ {
2173
+ "entropy": 3.450204849243164,
2174
+ "epoch": 5.592783505154639,
2175
+ "grad_norm": 71.5,
2176
+ "learning_rate": 7.072574456667258e-08,
2177
+ "loss": 3.0189,
2178
+ "mean_token_accuracy": 0.47769983410835265,
2179
+ "num_tokens": 68567.0,
2180
+ "step": 2170
2181
+ },
2182
+ {
2183
+ "entropy": 3.2766273736953737,
2184
+ "epoch": 5.618556701030927,
2185
+ "grad_norm": 83.5,
2186
+ "learning_rate": 6.214507263376801e-08,
2187
+ "loss": 2.881,
2188
+ "mean_token_accuracy": 0.48972640335559847,
2189
+ "num_tokens": 68846.0,
2190
+ "step": 2180
2191
+ },
2192
+ {
2193
+ "entropy": 3.267438220977783,
2194
+ "epoch": 5.644329896907217,
2195
+ "grad_norm": 68.0,
2196
+ "learning_rate": 5.411259108134115e-08,
2197
+ "loss": 2.9084,
2198
+ "mean_token_accuracy": 0.4877107352018356,
2199
+ "num_tokens": 69197.0,
2200
+ "step": 2190
2201
+ },
2202
+ {
2203
+ "entropy": 3.320011782646179,
2204
+ "epoch": 5.670103092783505,
2205
+ "grad_norm": 70.0,
2206
+ "learning_rate": 4.663010614038205e-08,
2207
+ "loss": 3.1586,
2208
+ "mean_token_accuracy": 0.4482086032629013,
2209
+ "num_tokens": 69493.0,
2210
+ "step": 2200
2211
+ },
2212
+ {
2213
+ "entropy": 3.4605057001113892,
2214
+ "epoch": 5.695876288659794,
2215
+ "grad_norm": 86.0,
2216
+ "learning_rate": 3.9699300366410895e-08,
2217
+ "loss": 3.2727,
2218
+ "mean_token_accuracy": 0.4599607527256012,
2219
+ "num_tokens": 69832.0,
2220
+ "step": 2210
2221
+ },
2222
+ {
2223
+ "entropy": 3.275946855545044,
2224
+ "epoch": 5.721649484536083,
2225
+ "grad_norm": 58.5,
2226
+ "learning_rate": 3.332173226113067e-08,
2227
+ "loss": 3.0561,
2228
+ "mean_token_accuracy": 0.460858029127121,
2229
+ "num_tokens": 70215.0,
2230
+ "step": 2220
2231
+ },
2232
+ {
2233
+ "entropy": 3.389675521850586,
2234
+ "epoch": 5.747422680412371,
2235
+ "grad_norm": 92.0,
2236
+ "learning_rate": 2.7498835921971058e-08,
2237
+ "loss": 3.0778,
2238
+ "mean_token_accuracy": 0.44814312756061553,
2239
+ "num_tokens": 70534.0,
2240
+ "step": 2230
2241
+ },
2242
+ {
2243
+ "entropy": 3.1975868225097654,
2244
+ "epoch": 5.77319587628866,
2245
+ "grad_norm": 72.5,
2246
+ "learning_rate": 2.2231920719610057e-08,
2247
+ "loss": 2.9487,
2248
+ "mean_token_accuracy": 0.47108798921108247,
2249
+ "num_tokens": 70838.0,
2250
+ "step": 2240
2251
+ },
2252
+ {
2253
+ "entropy": 3.20022566318512,
2254
+ "epoch": 5.798969072164948,
2255
+ "grad_norm": 69.0,
2256
+ "learning_rate": 1.75221710035417e-08,
2257
+ "loss": 2.67,
2258
+ "mean_token_accuracy": 0.5278927952051162,
2259
+ "num_tokens": 71133.0,
2260
+ "step": 2250
2261
+ },
2262
+ {
2263
+ "entropy": 3.337849473953247,
2264
+ "epoch": 5.824742268041237,
2265
+ "grad_norm": 93.5,
2266
+ "learning_rate": 1.3370645835754926e-08,
2267
+ "loss": 3.1954,
2268
+ "mean_token_accuracy": 0.4186277240514755,
2269
+ "num_tokens": 71475.0,
2270
+ "step": 2260
2271
+ },
2272
+ {
2273
+ "entropy": 3.2620397567749024,
2274
+ "epoch": 5.850515463917525,
2275
+ "grad_norm": 99.0,
2276
+ "learning_rate": 9.77827875258769e-09,
2277
+ "loss": 3.0062,
2278
+ "mean_token_accuracy": 0.4634517639875412,
2279
+ "num_tokens": 71766.0,
2280
+ "step": 2270
2281
+ },
2282
+ {
2283
+ "entropy": 3.139668273925781,
2284
+ "epoch": 5.876288659793815,
2285
+ "grad_norm": 84.0,
2286
+ "learning_rate": 6.745877554806268e-09,
2287
+ "loss": 3.0198,
2288
+ "mean_token_accuracy": 0.4661685138940811,
2289
+ "num_tokens": 72053.0,
2290
+ "step": 2280
2291
+ },
2292
+ {
2293
+ "entropy": 3.293707489967346,
2294
+ "epoch": 5.902061855670103,
2295
+ "grad_norm": 59.0,
2296
+ "learning_rate": 4.274124125958068e-09,
2297
+ "loss": 3.0505,
2298
+ "mean_token_accuracy": 0.46237342357635497,
2299
+ "num_tokens": 72347.0,
2300
+ "step": 2290
2301
+ },
2302
+ {
2303
+ "entropy": 3.4691567182540894,
2304
+ "epoch": 5.927835051546392,
2305
+ "grad_norm": 59.5,
2306
+ "learning_rate": 2.363574279040104e-09,
2307
+ "loss": 3.3964,
2308
+ "mean_token_accuracy": 0.4291213423013687,
2309
+ "num_tokens": 72681.0,
2310
+ "step": 2300
2311
+ }
2312
+ ],
2313
+ "logging_steps": 10,
2314
+ "max_steps": 2328,
2315
+ "num_input_tokens_seen": 0,
2316
+ "num_train_epochs": 6,
2317
+ "save_steps": 100,
2318
+ "stateful_callbacks": {
2319
+ "TrainerControl": {
2320
+ "args": {
2321
+ "should_epoch_stop": false,
2322
+ "should_evaluate": false,
2323
+ "should_log": false,
2324
+ "should_save": true,
2325
+ "should_training_stop": false
2326
+ },
2327
+ "attributes": {}
2328
+ }
2329
+ },
2330
+ "total_flos": 614623990192128.0,
2331
+ "train_batch_size": 1,
2332
+ "trial_name": null,
2333
+ "trial_params": null
2334
+ }
checkpoint-2300/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:609c0100e8e8ae1a8654246f76d53314d98a0e4dcf6191e136010cdaddbb8430
3
+ size 6225
checkpoint-2300/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-2328/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
checkpoint-2328/chat_template.jinja ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
27
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
28
+ {%- elif message.role == "assistant" %}
29
+ {%- set content = message.content %}
30
+ {%- set reasoning_content = '' %}
31
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
32
+ {%- set reasoning_content = message.reasoning_content %}
33
+ {%- else %}
34
+ {%- if '</think>' in message.content %}
35
+ {%- set content = message.content.split('</think>')[-1].lstrip('\n') %}
36
+ {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
37
+ {%- endif %}
38
+ {%- endif %}
39
+ {%- if loop.index0 > ns.last_query_index %}
40
+ {%- if loop.last or (not loop.last and reasoning_content) %}
41
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
42
+ {%- else %}
43
+ {{- '<|im_start|>' + message.role + '\n' + content }}
44
+ {%- endif %}
45
+ {%- else %}
46
+ {{- '<|im_start|>' + message.role + '\n' + content }}
47
+ {%- endif %}
48
+ {%- if message.tool_calls %}
49
+ {%- for tool_call in message.tool_calls %}
50
+ {%- if (loop.first and content) or (not loop.first) %}
51
+ {{- '\n' }}
52
+ {%- endif %}
53
+ {%- if tool_call.function %}
54
+ {%- set tool_call = tool_call.function %}
55
+ {%- endif %}
56
+ {{- '<tool_call>\n{"name": "' }}
57
+ {{- tool_call.name }}
58
+ {{- '", "arguments": ' }}
59
+ {%- if tool_call.arguments is string %}
60
+ {{- tool_call.arguments }}
61
+ {%- else %}
62
+ {{- tool_call.arguments | tojson }}
63
+ {%- endif %}
64
+ {{- '}\n</tool_call>' }}
65
+ {%- endfor %}
66
+ {%- endif %}
67
+ {{- '<|im_end|>\n' }}
68
+ {%- elif message.role == "tool" %}
69
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
70
+ {{- '<|im_start|>user' }}
71
+ {%- endif %}
72
+ {{- '\n<tool_response>\n' }}
73
+ {{- message.content }}
74
+ {{- '\n</tool_response>' }}
75
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
76
+ {{- '<|im_end|>\n' }}
77
+ {%- endif %}
78
+ {%- endif %}
79
+ {%- endfor %}
80
+ {%- if add_generation_prompt %}
81
+ {{- '<|im_start|>assistant\n' }}
82
+ {%- if enable_thinking is defined and enable_thinking is false %}
83
+ {{- '<think>\n\n</think>\n\n' }}
84
+ {%- endif %}
85
+ {%- endif %}
checkpoint-2328/config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "dtype": "bfloat16",
8
+ "eos_token_id": 151643,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 2048,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 6144,
14
+ "layer_types": [
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention"
43
+ ],
44
+ "max_position_embeddings": 32768,
45
+ "max_window_layers": 28,
46
+ "model_type": "qwen3",
47
+ "num_attention_heads": 16,
48
+ "num_hidden_layers": 28,
49
+ "num_key_value_heads": 8,
50
+ "pad_token_id": 151643,
51
+ "rms_norm_eps": 1e-06,
52
+ "rope_scaling": null,
53
+ "rope_theta": 1000000,
54
+ "sliding_window": null,
55
+ "tie_word_embeddings": true,
56
+ "transformers_version": "4.57.3",
57
+ "use_cache": true,
58
+ "use_sliding_window": false,
59
+ "vocab_size": 151936
60
+ }
checkpoint-2328/generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token_id": [
3
+ 151643
4
+ ],
5
+ "max_new_tokens": 2048,
6
+ "pad_token_id": 151643,
7
+ "transformers_version": "4.57.3"
8
+ }
checkpoint-2328/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-2328/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:899a58d2b285510f699ed77c114a40c8e6b949d639706ecc55e1074ca1ae1d00
3
+ size 3441185608
checkpoint-2328/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6340855080cd5327418de99e4f095ff546917fa44c7c86edb919663a3b60ece
3
+ size 6882567855
checkpoint-2328/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fd9432b50864b1799d071a2391f13a188cc959e985675ab69fd688672db2853
3
+ size 14645
checkpoint-2328/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7e47ebeeffa48060a6922bdf8fb1ef41a600a70bae118250e1ebd5e3f3280a6
3
+ size 1465
checkpoint-2328/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
checkpoint-2328/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
checkpoint-2328/tokenizer_config.json ADDED
@@ -0,0 +1,239 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|endoftext|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 131072,
235
+ "pad_token": "<|endoftext|>",
236
+ "split_special_tokens": false,
237
+ "tokenizer_class": "Qwen2Tokenizer",
238
+ "unk_token": null
239
+ }
checkpoint-2328/trainer_state.json ADDED
@@ -0,0 +1,2354 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 6.0,
6
+ "eval_steps": 500,
7
+ "global_step": 2328,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "entropy": 4.914482545852661,
14
+ "epoch": 0.02577319587628866,
15
+ "grad_norm": 232.0,
16
+ "learning_rate": 1.9313304721030046e-07,
17
+ "loss": 5.5997,
18
+ "mean_token_accuracy": 0.20726535096764565,
19
+ "num_tokens": 266.0,
20
+ "step": 10
21
+ },
22
+ {
23
+ "entropy": 4.538765811920166,
24
+ "epoch": 0.05154639175257732,
25
+ "grad_norm": 160.0,
26
+ "learning_rate": 4.07725321888412e-07,
27
+ "loss": 5.4934,
28
+ "mean_token_accuracy": 0.21887856498360633,
29
+ "num_tokens": 627.0,
30
+ "step": 20
31
+ },
32
+ {
33
+ "entropy": 4.939351224899292,
34
+ "epoch": 0.07731958762886598,
35
+ "grad_norm": 118.5,
36
+ "learning_rate": 6.223175965665236e-07,
37
+ "loss": 5.7523,
38
+ "mean_token_accuracy": 0.1917542487382889,
39
+ "num_tokens": 926.0,
40
+ "step": 30
41
+ },
42
+ {
43
+ "entropy": 4.698552560806275,
44
+ "epoch": 0.10309278350515463,
45
+ "grad_norm": 187.0,
46
+ "learning_rate": 8.369098712446352e-07,
47
+ "loss": 5.1204,
48
+ "mean_token_accuracy": 0.22331946194171906,
49
+ "num_tokens": 1259.0,
50
+ "step": 40
51
+ },
52
+ {
53
+ "entropy": 4.72123908996582,
54
+ "epoch": 0.12886597938144329,
55
+ "grad_norm": 166.0,
56
+ "learning_rate": 1.051502145922747e-06,
57
+ "loss": 5.198,
58
+ "mean_token_accuracy": 0.21775908395648003,
59
+ "num_tokens": 1550.0,
60
+ "step": 50
61
+ },
62
+ {
63
+ "entropy": 4.719539976119995,
64
+ "epoch": 0.15463917525773196,
65
+ "grad_norm": 200.0,
66
+ "learning_rate": 1.2660944206008586e-06,
67
+ "loss": 5.0832,
68
+ "mean_token_accuracy": 0.2405412092804909,
69
+ "num_tokens": 1839.0,
70
+ "step": 60
71
+ },
72
+ {
73
+ "entropy": 4.78384747505188,
74
+ "epoch": 0.18041237113402062,
75
+ "grad_norm": 149.0,
76
+ "learning_rate": 1.48068669527897e-06,
77
+ "loss": 5.0961,
78
+ "mean_token_accuracy": 0.19893446192145348,
79
+ "num_tokens": 2171.0,
80
+ "step": 70
81
+ },
82
+ {
83
+ "entropy": 4.974851608276367,
84
+ "epoch": 0.20618556701030927,
85
+ "grad_norm": 177.0,
86
+ "learning_rate": 1.6952789699570817e-06,
87
+ "loss": 5.2262,
88
+ "mean_token_accuracy": 0.20617640018463135,
89
+ "num_tokens": 2450.0,
90
+ "step": 80
91
+ },
92
+ {
93
+ "entropy": 4.705893039703369,
94
+ "epoch": 0.23195876288659795,
95
+ "grad_norm": 182.0,
96
+ "learning_rate": 1.9098712446351934e-06,
97
+ "loss": 4.8345,
98
+ "mean_token_accuracy": 0.22286981157958508,
99
+ "num_tokens": 2786.0,
100
+ "step": 90
101
+ },
102
+ {
103
+ "entropy": 4.696354675292969,
104
+ "epoch": 0.25773195876288657,
105
+ "grad_norm": 154.0,
106
+ "learning_rate": 2.124463519313305e-06,
107
+ "loss": 5.023,
108
+ "mean_token_accuracy": 0.25020611882209776,
109
+ "num_tokens": 3090.0,
110
+ "step": 100
111
+ },
112
+ {
113
+ "entropy": 4.626954746246338,
114
+ "epoch": 0.28350515463917525,
115
+ "grad_norm": 59.0,
116
+ "learning_rate": 2.3390557939914167e-06,
117
+ "loss": 4.5576,
118
+ "mean_token_accuracy": 0.2621785670518875,
119
+ "num_tokens": 3383.0,
120
+ "step": 110
121
+ },
122
+ {
123
+ "entropy": 4.682432174682617,
124
+ "epoch": 0.30927835051546393,
125
+ "grad_norm": 76.0,
126
+ "learning_rate": 2.553648068669528e-06,
127
+ "loss": 4.792,
128
+ "mean_token_accuracy": 0.26012246310710907,
129
+ "num_tokens": 3650.0,
130
+ "step": 120
131
+ },
132
+ {
133
+ "entropy": 4.563408708572387,
134
+ "epoch": 0.33505154639175255,
135
+ "grad_norm": 43.25,
136
+ "learning_rate": 2.7682403433476396e-06,
137
+ "loss": 4.5349,
138
+ "mean_token_accuracy": 0.22586977183818818,
139
+ "num_tokens": 3990.0,
140
+ "step": 130
141
+ },
142
+ {
143
+ "entropy": 4.270893955230713,
144
+ "epoch": 0.36082474226804123,
145
+ "grad_norm": 65.0,
146
+ "learning_rate": 2.982832618025751e-06,
147
+ "loss": 4.147,
148
+ "mean_token_accuracy": 0.28786555826663973,
149
+ "num_tokens": 4345.0,
150
+ "step": 140
151
+ },
152
+ {
153
+ "entropy": 4.0949385404586796,
154
+ "epoch": 0.3865979381443299,
155
+ "grad_norm": 55.75,
156
+ "learning_rate": 3.197424892703863e-06,
157
+ "loss": 3.7088,
158
+ "mean_token_accuracy": 0.34470676481723783,
159
+ "num_tokens": 4601.0,
160
+ "step": 150
161
+ },
162
+ {
163
+ "entropy": 4.128937864303589,
164
+ "epoch": 0.41237113402061853,
165
+ "grad_norm": 49.25,
166
+ "learning_rate": 3.412017167381975e-06,
167
+ "loss": 3.9797,
168
+ "mean_token_accuracy": 0.3258319616317749,
169
+ "num_tokens": 5024.0,
170
+ "step": 160
171
+ },
172
+ {
173
+ "entropy": 3.8451520442962646,
174
+ "epoch": 0.4381443298969072,
175
+ "grad_norm": 72.5,
176
+ "learning_rate": 3.6266094420600863e-06,
177
+ "loss": 3.7548,
178
+ "mean_token_accuracy": 0.3747887283563614,
179
+ "num_tokens": 5311.0,
180
+ "step": 170
181
+ },
182
+ {
183
+ "entropy": 3.712318539619446,
184
+ "epoch": 0.4639175257731959,
185
+ "grad_norm": 44.0,
186
+ "learning_rate": 3.841201716738197e-06,
187
+ "loss": 3.6765,
188
+ "mean_token_accuracy": 0.3758635461330414,
189
+ "num_tokens": 5651.0,
190
+ "step": 180
191
+ },
192
+ {
193
+ "entropy": 3.763609743118286,
194
+ "epoch": 0.4896907216494845,
195
+ "grad_norm": 43.25,
196
+ "learning_rate": 4.055793991416309e-06,
197
+ "loss": 3.6278,
198
+ "mean_token_accuracy": 0.3873393088579178,
199
+ "num_tokens": 5983.0,
200
+ "step": 190
201
+ },
202
+ {
203
+ "entropy": 3.781333661079407,
204
+ "epoch": 0.5154639175257731,
205
+ "grad_norm": 67.0,
206
+ "learning_rate": 4.270386266094421e-06,
207
+ "loss": 3.7018,
208
+ "mean_token_accuracy": 0.3747037798166275,
209
+ "num_tokens": 6296.0,
210
+ "step": 200
211
+ },
212
+ {
213
+ "entropy": 3.6542357921600344,
214
+ "epoch": 0.5412371134020618,
215
+ "grad_norm": 63.75,
216
+ "learning_rate": 4.484978540772533e-06,
217
+ "loss": 3.2866,
218
+ "mean_token_accuracy": 0.4405659481883049,
219
+ "num_tokens": 6658.0,
220
+ "step": 210
221
+ },
222
+ {
223
+ "entropy": 3.38469135761261,
224
+ "epoch": 0.5670103092783505,
225
+ "grad_norm": 88.0,
226
+ "learning_rate": 4.699570815450644e-06,
227
+ "loss": 3.1957,
228
+ "mean_token_accuracy": 0.41826934069395066,
229
+ "num_tokens": 6997.0,
230
+ "step": 220
231
+ },
232
+ {
233
+ "entropy": 3.9692420244216917,
234
+ "epoch": 0.5927835051546392,
235
+ "grad_norm": 42.5,
236
+ "learning_rate": 4.914163090128756e-06,
237
+ "loss": 4.0617,
238
+ "mean_token_accuracy": 0.3779368013143539,
239
+ "num_tokens": 7376.0,
240
+ "step": 230
241
+ },
242
+ {
243
+ "entropy": 3.725129175186157,
244
+ "epoch": 0.6185567010309279,
245
+ "grad_norm": 49.5,
246
+ "learning_rate": 4.999898809142829e-06,
247
+ "loss": 3.3293,
248
+ "mean_token_accuracy": 0.4062195152044296,
249
+ "num_tokens": 7688.0,
250
+ "step": 240
251
+ },
252
+ {
253
+ "entropy": 3.673329973220825,
254
+ "epoch": 0.6443298969072165,
255
+ "grad_norm": 84.5,
256
+ "learning_rate": 4.9992804502362914e-06,
257
+ "loss": 3.5039,
258
+ "mean_token_accuracy": 0.41747846007347106,
259
+ "num_tokens": 7988.0,
260
+ "step": 250
261
+ },
262
+ {
263
+ "entropy": 3.5831470012664797,
264
+ "epoch": 0.6701030927835051,
265
+ "grad_norm": 79.0,
266
+ "learning_rate": 4.998100088445351e-06,
267
+ "loss": 3.3134,
268
+ "mean_token_accuracy": 0.4633125364780426,
269
+ "num_tokens": 8263.0,
270
+ "step": 260
271
+ },
272
+ {
273
+ "entropy": 3.5552636861801146,
274
+ "epoch": 0.6958762886597938,
275
+ "grad_norm": 65.0,
276
+ "learning_rate": 4.996357989193094e-06,
277
+ "loss": 3.4462,
278
+ "mean_token_accuracy": 0.40612466633319855,
279
+ "num_tokens": 8590.0,
280
+ "step": 270
281
+ },
282
+ {
283
+ "entropy": 3.833337092399597,
284
+ "epoch": 0.7216494845360825,
285
+ "grad_norm": 39.5,
286
+ "learning_rate": 4.994054544218193e-06,
287
+ "loss": 3.7615,
288
+ "mean_token_accuracy": 0.3914962366223335,
289
+ "num_tokens": 8945.0,
290
+ "step": 280
291
+ },
292
+ {
293
+ "entropy": 3.6764284133911134,
294
+ "epoch": 0.7474226804123711,
295
+ "grad_norm": 60.25,
296
+ "learning_rate": 4.991190271486816e-06,
297
+ "loss": 3.4472,
298
+ "mean_token_accuracy": 0.4174343138933182,
299
+ "num_tokens": 9269.0,
300
+ "step": 290
301
+ },
302
+ {
303
+ "entropy": 3.5497223377227782,
304
+ "epoch": 0.7731958762886598,
305
+ "grad_norm": 76.0,
306
+ "learning_rate": 4.987765815076157e-06,
307
+ "loss": 3.1154,
308
+ "mean_token_accuracy": 0.44447133839130404,
309
+ "num_tokens": 9550.0,
310
+ "step": 300
311
+ },
312
+ {
313
+ "entropy": 3.8354516506195067,
314
+ "epoch": 0.7989690721649485,
315
+ "grad_norm": 58.25,
316
+ "learning_rate": 4.9837819450296e-06,
317
+ "loss": 3.4305,
318
+ "mean_token_accuracy": 0.42582335472106936,
319
+ "num_tokens": 9899.0,
320
+ "step": 310
321
+ },
322
+ {
323
+ "entropy": 3.752107357978821,
324
+ "epoch": 0.8247422680412371,
325
+ "grad_norm": 56.5,
326
+ "learning_rate": 4.979239557183571e-06,
327
+ "loss": 3.6763,
328
+ "mean_token_accuracy": 0.4198161542415619,
329
+ "num_tokens": 10174.0,
330
+ "step": 320
331
+ },
332
+ {
333
+ "entropy": 3.641463875770569,
334
+ "epoch": 0.8505154639175257,
335
+ "grad_norm": 61.25,
336
+ "learning_rate": 4.974139672966082e-06,
337
+ "loss": 3.5237,
338
+ "mean_token_accuracy": 0.3649302959442139,
339
+ "num_tokens": 10520.0,
340
+ "step": 330
341
+ },
342
+ {
343
+ "entropy": 3.487735724449158,
344
+ "epoch": 0.8762886597938144,
345
+ "grad_norm": 55.0,
346
+ "learning_rate": 4.968483439167061e-06,
347
+ "loss": 3.3832,
348
+ "mean_token_accuracy": 0.4202912449836731,
349
+ "num_tokens": 10834.0,
350
+ "step": 340
351
+ },
352
+ {
353
+ "entropy": 3.659078526496887,
354
+ "epoch": 0.9020618556701031,
355
+ "grad_norm": 58.0,
356
+ "learning_rate": 4.9622721276804674e-06,
357
+ "loss": 3.0981,
358
+ "mean_token_accuracy": 0.45194968581199646,
359
+ "num_tokens": 11110.0,
360
+ "step": 350
361
+ },
362
+ {
363
+ "entropy": 3.3279967069625855,
364
+ "epoch": 0.9278350515463918,
365
+ "grad_norm": 81.5,
366
+ "learning_rate": 4.955507135218291e-06,
367
+ "loss": 3.2825,
368
+ "mean_token_accuracy": 0.42670700550079343,
369
+ "num_tokens": 11410.0,
370
+ "step": 360
371
+ },
372
+ {
373
+ "entropy": 3.2990634202957154,
374
+ "epoch": 0.9536082474226805,
375
+ "grad_norm": 52.25,
376
+ "learning_rate": 4.948189982996479e-06,
377
+ "loss": 2.945,
378
+ "mean_token_accuracy": 0.46619434356689454,
379
+ "num_tokens": 11758.0,
380
+ "step": 370
381
+ },
382
+ {
383
+ "entropy": 3.442564105987549,
384
+ "epoch": 0.979381443298969,
385
+ "grad_norm": 77.0,
386
+ "learning_rate": 4.940322316392865e-06,
387
+ "loss": 3.4907,
388
+ "mean_token_accuracy": 0.465116411447525,
389
+ "num_tokens": 12008.0,
390
+ "step": 380
391
+ },
392
+ {
393
+ "entropy": 3.4102420568466187,
394
+ "epoch": 1.0051546391752577,
395
+ "grad_norm": 105.5,
396
+ "learning_rate": 4.931905904577182e-06,
397
+ "loss": 3.4858,
398
+ "mean_token_accuracy": 0.4123460859060287,
399
+ "num_tokens": 12288.0,
400
+ "step": 390
401
+ },
402
+ {
403
+ "entropy": 3.615120697021484,
404
+ "epoch": 1.0309278350515463,
405
+ "grad_norm": 57.75,
406
+ "learning_rate": 4.922942640113234e-06,
407
+ "loss": 3.1624,
408
+ "mean_token_accuracy": 0.45850674211978915,
409
+ "num_tokens": 12600.0,
410
+ "step": 400
411
+ },
412
+ {
413
+ "entropy": 3.5458458185195925,
414
+ "epoch": 1.056701030927835,
415
+ "grad_norm": 64.5,
416
+ "learning_rate": 4.913434538533324e-06,
417
+ "loss": 3.3334,
418
+ "mean_token_accuracy": 0.46345340013504027,
419
+ "num_tokens": 12888.0,
420
+ "step": 410
421
+ },
422
+ {
423
+ "entropy": 3.283875823020935,
424
+ "epoch": 1.0824742268041236,
425
+ "grad_norm": 84.5,
426
+ "learning_rate": 4.90338373788503e-06,
427
+ "loss": 3.2497,
428
+ "mean_token_accuracy": 0.47709590196609497,
429
+ "num_tokens": 13180.0,
430
+ "step": 420
431
+ },
432
+ {
433
+ "entropy": 3.4606318950653074,
434
+ "epoch": 1.1082474226804124,
435
+ "grad_norm": 72.0,
436
+ "learning_rate": 4.892792498250431e-06,
437
+ "loss": 3.5329,
438
+ "mean_token_accuracy": 0.4233353078365326,
439
+ "num_tokens": 13485.0,
440
+ "step": 430
441
+ },
442
+ {
443
+ "entropy": 3.3683727979660034,
444
+ "epoch": 1.134020618556701,
445
+ "grad_norm": 58.25,
446
+ "learning_rate": 4.881663201237889e-06,
447
+ "loss": 3.0689,
448
+ "mean_token_accuracy": 0.4563048958778381,
449
+ "num_tokens": 13849.0,
450
+ "step": 440
451
+ },
452
+ {
453
+ "entropy": 3.3520001411437987,
454
+ "epoch": 1.1597938144329896,
455
+ "grad_norm": 49.5,
456
+ "learning_rate": 4.869998349446514e-06,
457
+ "loss": 3.0789,
458
+ "mean_token_accuracy": 0.4614581674337387,
459
+ "num_tokens": 14121.0,
460
+ "step": 450
461
+ },
462
+ {
463
+ "entropy": 3.3299176692962646,
464
+ "epoch": 1.1855670103092784,
465
+ "grad_norm": 64.5,
466
+ "learning_rate": 4.857800565903405e-06,
467
+ "loss": 3.2253,
468
+ "mean_token_accuracy": 0.4312807470560074,
469
+ "num_tokens": 14503.0,
470
+ "step": 460
471
+ },
472
+ {
473
+ "entropy": 3.6635308504104613,
474
+ "epoch": 1.211340206185567,
475
+ "grad_norm": 52.5,
476
+ "learning_rate": 4.845072593473826e-06,
477
+ "loss": 3.2584,
478
+ "mean_token_accuracy": 0.46550854444503786,
479
+ "num_tokens": 14792.0,
480
+ "step": 470
481
+ },
482
+ {
483
+ "entropy": 3.2862602949142454,
484
+ "epoch": 1.2371134020618557,
485
+ "grad_norm": 56.25,
486
+ "learning_rate": 4.831817294244432e-06,
487
+ "loss": 3.1376,
488
+ "mean_token_accuracy": 0.4864483565092087,
489
+ "num_tokens": 15071.0,
490
+ "step": 480
491
+ },
492
+ {
493
+ "entropy": 3.4670738697052004,
494
+ "epoch": 1.2628865979381443,
495
+ "grad_norm": 60.25,
496
+ "learning_rate": 4.8180376488796755e-06,
497
+ "loss": 3.4598,
498
+ "mean_token_accuracy": 0.38327425718307495,
499
+ "num_tokens": 15340.0,
500
+ "step": 490
501
+ },
502
+ {
503
+ "entropy": 3.4762473344802856,
504
+ "epoch": 1.2886597938144329,
505
+ "grad_norm": 72.0,
506
+ "learning_rate": 4.803736755951564e-06,
507
+ "loss": 3.23,
508
+ "mean_token_accuracy": 0.4640804290771484,
509
+ "num_tokens": 15658.0,
510
+ "step": 500
511
+ },
512
+ {
513
+ "entropy": 3.4152861833572388,
514
+ "epoch": 1.3144329896907216,
515
+ "grad_norm": 60.5,
516
+ "learning_rate": 4.788917831242895e-06,
517
+ "loss": 3.1239,
518
+ "mean_token_accuracy": 0.47079411447048186,
519
+ "num_tokens": 15943.0,
520
+ "step": 510
521
+ },
522
+ {
523
+ "entropy": 3.3919612884521486,
524
+ "epoch": 1.3402061855670104,
525
+ "grad_norm": 116.5,
526
+ "learning_rate": 4.773584207024135e-06,
527
+ "loss": 3.193,
528
+ "mean_token_accuracy": 0.40636845529079435,
529
+ "num_tokens": 16268.0,
530
+ "step": 520
531
+ },
532
+ {
533
+ "entropy": 3.520834803581238,
534
+ "epoch": 1.365979381443299,
535
+ "grad_norm": 67.5,
536
+ "learning_rate": 4.7577393313041025e-06,
537
+ "loss": 3.3889,
538
+ "mean_token_accuracy": 0.4428008824586868,
539
+ "num_tokens": 16611.0,
540
+ "step": 530
541
+ },
542
+ {
543
+ "entropy": 3.5736474275588987,
544
+ "epoch": 1.3917525773195876,
545
+ "grad_norm": 51.25,
546
+ "learning_rate": 4.741386767054636e-06,
547
+ "loss": 3.4062,
548
+ "mean_token_accuracy": 0.408548378944397,
549
+ "num_tokens": 16917.0,
550
+ "step": 540
551
+ },
552
+ {
553
+ "entropy": 3.4252942323684694,
554
+ "epoch": 1.4175257731958764,
555
+ "grad_norm": 66.5,
556
+ "learning_rate": 4.724530191409399e-06,
557
+ "loss": 3.2558,
558
+ "mean_token_accuracy": 0.4369684547185898,
559
+ "num_tokens": 17283.0,
560
+ "step": 550
561
+ },
562
+ {
563
+ "entropy": 3.335077738761902,
564
+ "epoch": 1.443298969072165,
565
+ "grad_norm": 57.25,
566
+ "learning_rate": 4.707173394837017e-06,
567
+ "loss": 3.157,
568
+ "mean_token_accuracy": 0.4764990329742432,
569
+ "num_tokens": 17604.0,
570
+ "step": 560
571
+ },
572
+ {
573
+ "entropy": 3.519501209259033,
574
+ "epoch": 1.4690721649484537,
575
+ "grad_norm": 57.5,
576
+ "learning_rate": 4.689320280288731e-06,
577
+ "loss": 3.4428,
578
+ "mean_token_accuracy": 0.4320195406675339,
579
+ "num_tokens": 17939.0,
580
+ "step": 570
581
+ },
582
+ {
583
+ "entropy": 3.5304367780685424,
584
+ "epoch": 1.4948453608247423,
585
+ "grad_norm": 55.0,
586
+ "learning_rate": 4.67097486232076e-06,
587
+ "loss": 3.2563,
588
+ "mean_token_accuracy": 0.4069958388805389,
589
+ "num_tokens": 18234.0,
590
+ "step": 580
591
+ },
592
+ {
593
+ "entropy": 3.662277579307556,
594
+ "epoch": 1.5206185567010309,
595
+ "grad_norm": 67.0,
596
+ "learning_rate": 4.65214126619156e-06,
597
+ "loss": 3.8052,
598
+ "mean_token_accuracy": 0.40810766220092776,
599
+ "num_tokens": 18574.0,
600
+ "step": 590
601
+ },
602
+ {
603
+ "entropy": 3.2554187536239625,
604
+ "epoch": 1.5463917525773194,
605
+ "grad_norm": 69.0,
606
+ "learning_rate": 4.632823726934199e-06,
607
+ "loss": 3.0088,
608
+ "mean_token_accuracy": 0.4515380173921585,
609
+ "num_tokens": 18886.0,
610
+ "step": 600
611
+ },
612
+ {
613
+ "entropy": 3.4018918752670286,
614
+ "epoch": 1.5721649484536082,
615
+ "grad_norm": 60.25,
616
+ "learning_rate": 4.613026588404036e-06,
617
+ "loss": 3.3767,
618
+ "mean_token_accuracy": 0.4256245791912079,
619
+ "num_tokens": 19227.0,
620
+ "step": 610
621
+ },
622
+ {
623
+ "entropy": 3.4922634601593017,
624
+ "epoch": 1.597938144329897,
625
+ "grad_norm": 45.5,
626
+ "learning_rate": 4.592754302301942e-06,
627
+ "loss": 3.0169,
628
+ "mean_token_accuracy": 0.4801509857177734,
629
+ "num_tokens": 19582.0,
630
+ "step": 620
631
+ },
632
+ {
633
+ "entropy": 3.3631627798080443,
634
+ "epoch": 1.6237113402061856,
635
+ "grad_norm": 74.0,
636
+ "learning_rate": 4.572011427173263e-06,
637
+ "loss": 3.2208,
638
+ "mean_token_accuracy": 0.4472234547138214,
639
+ "num_tokens": 19934.0,
640
+ "step": 630
641
+ },
642
+ {
643
+ "entropy": 3.353942036628723,
644
+ "epoch": 1.6494845360824741,
645
+ "grad_norm": 54.25,
646
+ "learning_rate": 4.550802627382756e-06,
647
+ "loss": 3.2015,
648
+ "mean_token_accuracy": 0.4314853399991989,
649
+ "num_tokens": 20286.0,
650
+ "step": 640
651
+ },
652
+ {
653
+ "entropy": 3.338911604881287,
654
+ "epoch": 1.675257731958763,
655
+ "grad_norm": 89.0,
656
+ "learning_rate": 4.529132672065738e-06,
657
+ "loss": 3.09,
658
+ "mean_token_accuracy": 0.46670873165130616,
659
+ "num_tokens": 20556.0,
660
+ "step": 650
661
+ },
662
+ {
663
+ "entropy": 3.380303430557251,
664
+ "epoch": 1.7010309278350515,
665
+ "grad_norm": 76.0,
666
+ "learning_rate": 4.507006434055663e-06,
667
+ "loss": 3.2405,
668
+ "mean_token_accuracy": 0.4376333147287369,
669
+ "num_tokens": 20862.0,
670
+ "step": 660
671
+ },
672
+ {
673
+ "entropy": 3.491442632675171,
674
+ "epoch": 1.7268041237113403,
675
+ "grad_norm": 50.25,
676
+ "learning_rate": 4.484428888788395e-06,
677
+ "loss": 3.4112,
678
+ "mean_token_accuracy": 0.43711408972740173,
679
+ "num_tokens": 21164.0,
680
+ "step": 670
681
+ },
682
+ {
683
+ "entropy": 3.451434350013733,
684
+ "epoch": 1.7525773195876289,
685
+ "grad_norm": 105.5,
686
+ "learning_rate": 4.461405113183396e-06,
687
+ "loss": 3.0601,
688
+ "mean_token_accuracy": 0.42957675755023955,
689
+ "num_tokens": 21460.0,
690
+ "step": 680
691
+ },
692
+ {
693
+ "entropy": 3.329232668876648,
694
+ "epoch": 1.7783505154639174,
695
+ "grad_norm": 63.25,
696
+ "learning_rate": 4.437940284502105e-06,
697
+ "loss": 3.1711,
698
+ "mean_token_accuracy": 0.4769585371017456,
699
+ "num_tokens": 21817.0,
700
+ "step": 690
701
+ },
702
+ {
703
+ "entropy": 3.1766634225845336,
704
+ "epoch": 1.8041237113402062,
705
+ "grad_norm": 47.25,
706
+ "learning_rate": 4.414039679183749e-06,
707
+ "loss": 3.5047,
708
+ "mean_token_accuracy": 0.43050127625465395,
709
+ "num_tokens": 22163.0,
710
+ "step": 700
711
+ },
712
+ {
713
+ "entropy": 3.394679617881775,
714
+ "epoch": 1.829896907216495,
715
+ "grad_norm": 64.0,
716
+ "learning_rate": 4.389708671658844e-06,
717
+ "loss": 3.0371,
718
+ "mean_token_accuracy": 0.45515852570533755,
719
+ "num_tokens": 22436.0,
720
+ "step": 710
721
+ },
722
+ {
723
+ "entropy": 3.4196753025054933,
724
+ "epoch": 1.8556701030927836,
725
+ "grad_norm": 217.0,
726
+ "learning_rate": 4.3649527331406796e-06,
727
+ "loss": 3.612,
728
+ "mean_token_accuracy": 0.4677813768386841,
729
+ "num_tokens": 22709.0,
730
+ "step": 720
731
+ },
732
+ {
733
+ "entropy": 3.4829108476638795,
734
+ "epoch": 1.8814432989690721,
735
+ "grad_norm": 60.25,
736
+ "learning_rate": 4.339777430395022e-06,
737
+ "loss": 3.3854,
738
+ "mean_token_accuracy": 0.41964206099510193,
739
+ "num_tokens": 23076.0,
740
+ "step": 730
741
+ },
742
+ {
743
+ "entropy": 3.3097528219223022,
744
+ "epoch": 1.9072164948453607,
745
+ "grad_norm": 63.25,
746
+ "learning_rate": 4.314188424488344e-06,
747
+ "loss": 2.8523,
748
+ "mean_token_accuracy": 0.48268236219882965,
749
+ "num_tokens": 23403.0,
750
+ "step": 740
751
+ },
752
+ {
753
+ "entropy": 3.3857040882110594,
754
+ "epoch": 1.9329896907216495,
755
+ "grad_norm": 65.5,
756
+ "learning_rate": 4.288191469514839e-06,
757
+ "loss": 2.9944,
758
+ "mean_token_accuracy": 0.43578424155712125,
759
+ "num_tokens": 23722.0,
760
+ "step": 750
761
+ },
762
+ {
763
+ "entropy": 3.408646750450134,
764
+ "epoch": 1.9587628865979383,
765
+ "grad_norm": 68.5,
766
+ "learning_rate": 4.261792411302525e-06,
767
+ "loss": 3.4775,
768
+ "mean_token_accuracy": 0.43519311845302583,
769
+ "num_tokens": 23993.0,
770
+ "step": 760
771
+ },
772
+ {
773
+ "entropy": 3.117679166793823,
774
+ "epoch": 1.9845360824742269,
775
+ "grad_norm": 57.25,
776
+ "learning_rate": 4.234997186098716e-06,
777
+ "loss": 2.829,
778
+ "mean_token_accuracy": 0.43563964366912844,
779
+ "num_tokens": 24324.0,
780
+ "step": 770
781
+ },
782
+ {
783
+ "entropy": 3.5846817255020142,
784
+ "epoch": 2.0103092783505154,
785
+ "grad_norm": 41.5,
786
+ "learning_rate": 4.207811819235164e-06,
787
+ "loss": 3.374,
788
+ "mean_token_accuracy": 0.4384703665971756,
789
+ "num_tokens": 24690.0,
790
+ "step": 780
791
+ },
792
+ {
793
+ "entropy": 3.0135667085647584,
794
+ "epoch": 2.036082474226804,
795
+ "grad_norm": 49.25,
796
+ "learning_rate": 4.180242423773166e-06,
797
+ "loss": 2.9591,
798
+ "mean_token_accuracy": 0.5096089750528335,
799
+ "num_tokens": 24959.0,
800
+ "step": 790
801
+ },
802
+ {
803
+ "entropy": 3.289955735206604,
804
+ "epoch": 2.0618556701030926,
805
+ "grad_norm": 52.75,
806
+ "learning_rate": 4.1522951991289465e-06,
807
+ "loss": 3.218,
808
+ "mean_token_accuracy": 0.39164299368858335,
809
+ "num_tokens": 25285.0,
810
+ "step": 800
811
+ },
812
+ {
813
+ "entropy": 3.3328727006912233,
814
+ "epoch": 2.0876288659793816,
815
+ "grad_norm": 58.25,
816
+ "learning_rate": 4.1239764296796175e-06,
817
+ "loss": 2.9698,
818
+ "mean_token_accuracy": 0.4740764260292053,
819
+ "num_tokens": 25578.0,
820
+ "step": 810
821
+ },
822
+ {
823
+ "entropy": 3.3463461637496947,
824
+ "epoch": 2.11340206185567,
825
+ "grad_norm": 74.0,
826
+ "learning_rate": 4.095292483350041e-06,
827
+ "loss": 3.2668,
828
+ "mean_token_accuracy": 0.41175087094306945,
829
+ "num_tokens": 25954.0,
830
+ "step": 820
831
+ },
832
+ {
833
+ "entropy": 3.347863268852234,
834
+ "epoch": 2.1391752577319587,
835
+ "grad_norm": 93.5,
836
+ "learning_rate": 4.066249810180895e-06,
837
+ "loss": 3.1272,
838
+ "mean_token_accuracy": 0.44678041338920593,
839
+ "num_tokens": 26225.0,
840
+ "step": 830
841
+ },
842
+ {
843
+ "entropy": 3.2035139322280886,
844
+ "epoch": 2.1649484536082473,
845
+ "grad_norm": 69.0,
846
+ "learning_rate": 4.036854940878284e-06,
847
+ "loss": 3.2895,
848
+ "mean_token_accuracy": 0.41500436663627627,
849
+ "num_tokens": 26615.0,
850
+ "step": 840
851
+ },
852
+ {
853
+ "entropy": 3.2597701787948608,
854
+ "epoch": 2.1907216494845363,
855
+ "grad_norm": 49.0,
856
+ "learning_rate": 4.007114485345205e-06,
857
+ "loss": 3.0027,
858
+ "mean_token_accuracy": 0.529634228348732,
859
+ "num_tokens": 26948.0,
860
+ "step": 850
861
+ },
862
+ {
863
+ "entropy": 3.3507386445999146,
864
+ "epoch": 2.216494845360825,
865
+ "grad_norm": 54.0,
866
+ "learning_rate": 3.977035131195202e-06,
867
+ "loss": 3.0962,
868
+ "mean_token_accuracy": 0.4740526854991913,
869
+ "num_tokens": 27268.0,
870
+ "step": 860
871
+ },
872
+ {
873
+ "entropy": 3.2031197786331176,
874
+ "epoch": 2.2422680412371134,
875
+ "grad_norm": 81.5,
876
+ "learning_rate": 3.946623642248554e-06,
877
+ "loss": 2.5841,
878
+ "mean_token_accuracy": 0.5312987476587295,
879
+ "num_tokens": 27549.0,
880
+ "step": 870
881
+ },
882
+ {
883
+ "entropy": 3.2550023555755616,
884
+ "epoch": 2.268041237113402,
885
+ "grad_norm": 69.5,
886
+ "learning_rate": 3.915886857011323e-06,
887
+ "loss": 3.1698,
888
+ "mean_token_accuracy": 0.4532750606536865,
889
+ "num_tokens": 27886.0,
890
+ "step": 880
891
+ },
892
+ {
893
+ "entropy": 3.485032653808594,
894
+ "epoch": 2.2938144329896906,
895
+ "grad_norm": 55.5,
896
+ "learning_rate": 3.8848316871376055e-06,
897
+ "loss": 3.4646,
898
+ "mean_token_accuracy": 0.4280081331729889,
899
+ "num_tokens": 28200.0,
900
+ "step": 890
901
+ },
902
+ {
903
+ "entropy": 3.3659554958343505,
904
+ "epoch": 2.319587628865979,
905
+ "grad_norm": 97.5,
906
+ "learning_rate": 3.853465115875335e-06,
907
+ "loss": 3.2682,
908
+ "mean_token_accuracy": 0.45062295794487,
909
+ "num_tokens": 28536.0,
910
+ "step": 900
911
+ },
912
+ {
913
+ "entropy": 3.13352746963501,
914
+ "epoch": 2.345360824742268,
915
+ "grad_norm": 80.0,
916
+ "learning_rate": 3.821794196495995e-06,
917
+ "loss": 2.9256,
918
+ "mean_token_accuracy": 0.4756322205066681,
919
+ "num_tokens": 28810.0,
920
+ "step": 910
921
+ },
922
+ {
923
+ "entropy": 3.135547161102295,
924
+ "epoch": 2.3711340206185567,
925
+ "grad_norm": 65.5,
926
+ "learning_rate": 3.7898260507085697e-06,
927
+ "loss": 2.9811,
928
+ "mean_token_accuracy": 0.4623017519712448,
929
+ "num_tokens": 29089.0,
930
+ "step": 920
931
+ },
932
+ {
933
+ "entropy": 3.449264574050903,
934
+ "epoch": 2.3969072164948453,
935
+ "grad_norm": 61.75,
936
+ "learning_rate": 3.757567867058125e-06,
937
+ "loss": 3.6536,
938
+ "mean_token_accuracy": 0.40380783975124357,
939
+ "num_tokens": 29403.0,
940
+ "step": 930
941
+ },
942
+ {
943
+ "entropy": 3.4234744548797607,
944
+ "epoch": 2.422680412371134,
945
+ "grad_norm": 61.5,
946
+ "learning_rate": 3.7250268993093396e-06,
947
+ "loss": 3.2193,
948
+ "mean_token_accuracy": 0.49216817915439603,
949
+ "num_tokens": 29689.0,
950
+ "step": 940
951
+ },
952
+ {
953
+ "entropy": 3.4770421504974367,
954
+ "epoch": 2.448453608247423,
955
+ "grad_norm": 44.0,
956
+ "learning_rate": 3.6922104648153885e-06,
957
+ "loss": 3.3501,
958
+ "mean_token_accuracy": 0.4592595547437668,
959
+ "num_tokens": 30047.0,
960
+ "step": 950
961
+ },
962
+ {
963
+ "entropy": 3.339312434196472,
964
+ "epoch": 2.4742268041237114,
965
+ "grad_norm": 55.0,
966
+ "learning_rate": 3.659125942872516e-06,
967
+ "loss": 3.075,
968
+ "mean_token_accuracy": 0.4719044387340546,
969
+ "num_tokens": 30376.0,
970
+ "step": 960
971
+ },
972
+ {
973
+ "entropy": 3.4688915252685546,
974
+ "epoch": 2.5,
975
+ "grad_norm": 61.75,
976
+ "learning_rate": 3.625780773060687e-06,
977
+ "loss": 3.1916,
978
+ "mean_token_accuracy": 0.41194990575313567,
979
+ "num_tokens": 30691.0,
980
+ "step": 970
981
+ },
982
+ {
983
+ "entropy": 3.302336239814758,
984
+ "epoch": 2.5257731958762886,
985
+ "grad_norm": 67.5,
986
+ "learning_rate": 3.5921824535706756e-06,
987
+ "loss": 3.1441,
988
+ "mean_token_accuracy": 0.4681535869836807,
989
+ "num_tokens": 31007.0,
990
+ "step": 980
991
+ },
992
+ {
993
+ "entropy": 3.343183135986328,
994
+ "epoch": 2.551546391752577,
995
+ "grad_norm": 71.0,
996
+ "learning_rate": 3.5583385395179793e-06,
997
+ "loss": 3.1333,
998
+ "mean_token_accuracy": 0.47766990661621095,
999
+ "num_tokens": 31272.0,
1000
+ "step": 990
1001
+ },
1002
+ {
1003
+ "entropy": 3.13364782333374,
1004
+ "epoch": 2.5773195876288657,
1005
+ "grad_norm": 80.5,
1006
+ "learning_rate": 3.5242566412439332e-06,
1007
+ "loss": 3.1096,
1008
+ "mean_token_accuracy": 0.46111657917499543,
1009
+ "num_tokens": 31549.0,
1010
+ "step": 1000
1011
+ },
1012
+ {
1013
+ "entropy": 3.2046863794326783,
1014
+ "epoch": 2.6030927835051547,
1015
+ "grad_norm": 103.0,
1016
+ "learning_rate": 3.4899444226044023e-06,
1017
+ "loss": 3.5947,
1018
+ "mean_token_accuracy": 0.41809163987636566,
1019
+ "num_tokens": 31839.0,
1020
+ "step": 1010
1021
+ },
1022
+ {
1023
+ "entropy": 3.2269564390182497,
1024
+ "epoch": 2.6288659793814433,
1025
+ "grad_norm": 60.0,
1026
+ "learning_rate": 3.455409599246442e-06,
1027
+ "loss": 3.0418,
1028
+ "mean_token_accuracy": 0.4730542838573456,
1029
+ "num_tokens": 32131.0,
1030
+ "step": 1020
1031
+ },
1032
+ {
1033
+ "entropy": 3.255638074874878,
1034
+ "epoch": 2.654639175257732,
1035
+ "grad_norm": 53.25,
1036
+ "learning_rate": 3.4206599368733114e-06,
1037
+ "loss": 3.1486,
1038
+ "mean_token_accuracy": 0.48221423029899596,
1039
+ "num_tokens": 32472.0,
1040
+ "step": 1030
1041
+ },
1042
+ {
1043
+ "entropy": 3.304290223121643,
1044
+ "epoch": 2.680412371134021,
1045
+ "grad_norm": 51.5,
1046
+ "learning_rate": 3.3857032494982327e-06,
1047
+ "loss": 3.1768,
1048
+ "mean_token_accuracy": 0.44427731931209563,
1049
+ "num_tokens": 32882.0,
1050
+ "step": 1040
1051
+ },
1052
+ {
1053
+ "entropy": 3.4136642932891847,
1054
+ "epoch": 2.7061855670103094,
1055
+ "grad_norm": 53.25,
1056
+ "learning_rate": 3.3505473976872883e-06,
1057
+ "loss": 3.1119,
1058
+ "mean_token_accuracy": 0.45050349533557893,
1059
+ "num_tokens": 33165.0,
1060
+ "step": 1050
1061
+ },
1062
+ {
1063
+ "entropy": 3.386446309089661,
1064
+ "epoch": 2.731958762886598,
1065
+ "grad_norm": 71.0,
1066
+ "learning_rate": 3.3152002867918433e-06,
1067
+ "loss": 3.2298,
1068
+ "mean_token_accuracy": 0.41270871758461,
1069
+ "num_tokens": 33441.0,
1070
+ "step": 1060
1071
+ },
1072
+ {
1073
+ "entropy": 3.571580958366394,
1074
+ "epoch": 2.7577319587628866,
1075
+ "grad_norm": 59.5,
1076
+ "learning_rate": 3.279669865170906e-06,
1077
+ "loss": 3.3263,
1078
+ "mean_token_accuracy": 0.45758517682552335,
1079
+ "num_tokens": 33761.0,
1080
+ "step": 1070
1081
+ },
1082
+ {
1083
+ "entropy": 3.489269471168518,
1084
+ "epoch": 2.783505154639175,
1085
+ "grad_norm": 78.0,
1086
+ "learning_rate": 3.2439641224038093e-06,
1087
+ "loss": 3.0612,
1088
+ "mean_token_accuracy": 0.4716973781585693,
1089
+ "num_tokens": 34086.0,
1090
+ "step": 1080
1091
+ },
1092
+ {
1093
+ "entropy": 3.352162575721741,
1094
+ "epoch": 2.8092783505154637,
1095
+ "grad_norm": 56.0,
1096
+ "learning_rate": 3.20809108749363e-06,
1097
+ "loss": 3.1394,
1098
+ "mean_token_accuracy": 0.45618820786476133,
1099
+ "num_tokens": 34422.0,
1100
+ "step": 1090
1101
+ },
1102
+ {
1103
+ "entropy": 3.4569438695907593,
1104
+ "epoch": 2.8350515463917527,
1105
+ "grad_norm": 76.5,
1106
+ "learning_rate": 3.17205882706174e-06,
1107
+ "loss": 3.0988,
1108
+ "mean_token_accuracy": 0.48512459397315977,
1109
+ "num_tokens": 34726.0,
1110
+ "step": 1100
1111
+ },
1112
+ {
1113
+ "entropy": 3.376886796951294,
1114
+ "epoch": 2.8608247422680413,
1115
+ "grad_norm": 103.0,
1116
+ "learning_rate": 3.135875443533896e-06,
1117
+ "loss": 3.4308,
1118
+ "mean_token_accuracy": 0.41293925344944,
1119
+ "num_tokens": 35057.0,
1120
+ "step": 1110
1121
+ },
1122
+ {
1123
+ "entropy": 3.207944130897522,
1124
+ "epoch": 2.88659793814433,
1125
+ "grad_norm": 72.5,
1126
+ "learning_rate": 3.0995490733182825e-06,
1127
+ "loss": 2.7498,
1128
+ "mean_token_accuracy": 0.47106904685497286,
1129
+ "num_tokens": 35327.0,
1130
+ "step": 1120
1131
+ },
1132
+ {
1133
+ "entropy": 3.2526353359222413,
1134
+ "epoch": 2.9123711340206184,
1135
+ "grad_norm": 73.5,
1136
+ "learning_rate": 3.06308788497591e-06,
1137
+ "loss": 3.1798,
1138
+ "mean_token_accuracy": 0.4563720256090164,
1139
+ "num_tokens": 35617.0,
1140
+ "step": 1130
1141
+ },
1142
+ {
1143
+ "entropy": 3.3556177854537963,
1144
+ "epoch": 2.9381443298969074,
1145
+ "grad_norm": 67.0,
1146
+ "learning_rate": 3.026500077383785e-06,
1147
+ "loss": 3.1254,
1148
+ "mean_token_accuracy": 0.45203186869621276,
1149
+ "num_tokens": 35941.0,
1150
+ "step": 1140
1151
+ },
1152
+ {
1153
+ "entropy": 3.4933328866958617,
1154
+ "epoch": 2.963917525773196,
1155
+ "grad_norm": 65.5,
1156
+ "learning_rate": 2.989793877891263e-06,
1157
+ "loss": 3.5153,
1158
+ "mean_token_accuracy": 0.38476662933826444,
1159
+ "num_tokens": 36320.0,
1160
+ "step": 1150
1161
+ },
1162
+ {
1163
+ "entropy": 3.307176351547241,
1164
+ "epoch": 2.9896907216494846,
1165
+ "grad_norm": 80.5,
1166
+ "learning_rate": 2.952977540469999e-06,
1167
+ "loss": 3.0624,
1168
+ "mean_token_accuracy": 0.4267964720726013,
1169
+ "num_tokens": 36595.0,
1170
+ "step": 1160
1171
+ },
1172
+ {
1173
+ "entropy": 3.25522928237915,
1174
+ "epoch": 3.015463917525773,
1175
+ "grad_norm": 71.5,
1176
+ "learning_rate": 2.9160593438579054e-06,
1177
+ "loss": 2.8878,
1178
+ "mean_token_accuracy": 0.4728631258010864,
1179
+ "num_tokens": 36963.0,
1180
+ "step": 1170
1181
+ },
1182
+ {
1183
+ "entropy": 3.160351610183716,
1184
+ "epoch": 3.0412371134020617,
1185
+ "grad_norm": 73.5,
1186
+ "learning_rate": 2.879047589697555e-06,
1187
+ "loss": 3.0562,
1188
+ "mean_token_accuracy": 0.5218929082155228,
1189
+ "num_tokens": 37236.0,
1190
+ "step": 1180
1191
+ },
1192
+ {
1193
+ "entropy": 3.497001218795776,
1194
+ "epoch": 3.0670103092783507,
1195
+ "grad_norm": 91.0,
1196
+ "learning_rate": 2.8419506006694186e-06,
1197
+ "loss": 3.2947,
1198
+ "mean_token_accuracy": 0.40737067759037016,
1199
+ "num_tokens": 37522.0,
1200
+ "step": 1190
1201
+ },
1202
+ {
1203
+ "entropy": 3.3410105228424074,
1204
+ "epoch": 3.0927835051546393,
1205
+ "grad_norm": 61.5,
1206
+ "learning_rate": 2.8047767186203808e-06,
1207
+ "loss": 3.2214,
1208
+ "mean_token_accuracy": 0.45971156358718873,
1209
+ "num_tokens": 37825.0,
1210
+ "step": 1200
1211
+ },
1212
+ {
1213
+ "entropy": 3.4304397821426393,
1214
+ "epoch": 3.118556701030928,
1215
+ "grad_norm": 68.0,
1216
+ "learning_rate": 2.767534302687942e-06,
1217
+ "loss": 3.5544,
1218
+ "mean_token_accuracy": 0.42803180813789365,
1219
+ "num_tokens": 38181.0,
1220
+ "step": 1210
1221
+ },
1222
+ {
1223
+ "entropy": 3.2708693027496336,
1224
+ "epoch": 3.1443298969072164,
1225
+ "grad_norm": 51.5,
1226
+ "learning_rate": 2.730231727420533e-06,
1227
+ "loss": 3.0771,
1228
+ "mean_token_accuracy": 0.42246846556663514,
1229
+ "num_tokens": 38521.0,
1230
+ "step": 1220
1231
+ },
1232
+ {
1233
+ "entropy": 3.278656005859375,
1234
+ "epoch": 3.170103092783505,
1235
+ "grad_norm": 58.75,
1236
+ "learning_rate": 2.6928773808943696e-06,
1237
+ "loss": 3.2341,
1238
+ "mean_token_accuracy": 0.4356242328882217,
1239
+ "num_tokens": 38806.0,
1240
+ "step": 1230
1241
+ },
1242
+ {
1243
+ "entropy": 3.4282007455825805,
1244
+ "epoch": 3.195876288659794,
1245
+ "grad_norm": 85.5,
1246
+ "learning_rate": 2.6554796628272534e-06,
1247
+ "loss": 3.3791,
1248
+ "mean_token_accuracy": 0.44304960370063784,
1249
+ "num_tokens": 39104.0,
1250
+ "step": 1240
1251
+ },
1252
+ {
1253
+ "entropy": 3.3579065799713135,
1254
+ "epoch": 3.2216494845360826,
1255
+ "grad_norm": 68.5,
1256
+ "learning_rate": 2.6180469826897683e-06,
1257
+ "loss": 3.3142,
1258
+ "mean_token_accuracy": 0.41488229632377627,
1259
+ "num_tokens": 39433.0,
1260
+ "step": 1250
1261
+ },
1262
+ {
1263
+ "entropy": 3.2836158752441404,
1264
+ "epoch": 3.247422680412371,
1265
+ "grad_norm": 58.25,
1266
+ "learning_rate": 2.5805877578142713e-06,
1267
+ "loss": 3.0095,
1268
+ "mean_token_accuracy": 0.47964567244052886,
1269
+ "num_tokens": 39730.0,
1270
+ "step": 1260
1271
+ },
1272
+ {
1273
+ "entropy": 3.3673797607421876,
1274
+ "epoch": 3.2731958762886597,
1275
+ "grad_norm": 76.0,
1276
+ "learning_rate": 2.543110411502129e-06,
1277
+ "loss": 2.7924,
1278
+ "mean_token_accuracy": 0.46048580706119535,
1279
+ "num_tokens": 40029.0,
1280
+ "step": 1270
1281
+ },
1282
+ {
1283
+ "entropy": 3.2721627712249757,
1284
+ "epoch": 3.2989690721649483,
1285
+ "grad_norm": 64.5,
1286
+ "learning_rate": 2.5056233711295985e-06,
1287
+ "loss": 3.0903,
1288
+ "mean_token_accuracy": 0.4774274632334709,
1289
+ "num_tokens": 40310.0,
1290
+ "step": 1280
1291
+ },
1292
+ {
1293
+ "entropy": 3.1327761888504027,
1294
+ "epoch": 3.3247422680412373,
1295
+ "grad_norm": 78.0,
1296
+ "learning_rate": 2.4681350662528004e-06,
1297
+ "loss": 2.9106,
1298
+ "mean_token_accuracy": 0.45230706930160525,
1299
+ "num_tokens": 40609.0,
1300
+ "step": 1290
1301
+ },
1302
+ {
1303
+ "entropy": 3.2892409801483153,
1304
+ "epoch": 3.350515463917526,
1305
+ "grad_norm": 89.5,
1306
+ "learning_rate": 2.4306539267122e-06,
1307
+ "loss": 3.2275,
1308
+ "mean_token_accuracy": 0.42299139201641084,
1309
+ "num_tokens": 40899.0,
1310
+ "step": 1300
1311
+ },
1312
+ {
1313
+ "entropy": 3.4429001808166504,
1314
+ "epoch": 3.3762886597938144,
1315
+ "grad_norm": 55.5,
1316
+ "learning_rate": 2.393188380737021e-06,
1317
+ "loss": 3.1632,
1318
+ "mean_token_accuracy": 0.45537458956241605,
1319
+ "num_tokens": 41229.0,
1320
+ "step": 1310
1321
+ },
1322
+ {
1323
+ "entropy": 3.1667943239212035,
1324
+ "epoch": 3.402061855670103,
1325
+ "grad_norm": 61.25,
1326
+ "learning_rate": 2.3557468530500298e-06,
1327
+ "loss": 3.0767,
1328
+ "mean_token_accuracy": 0.41762115657329557,
1329
+ "num_tokens": 41595.0,
1330
+ "step": 1320
1331
+ },
1332
+ {
1333
+ "entropy": 2.933236765861511,
1334
+ "epoch": 3.4278350515463916,
1335
+ "grad_norm": 51.0,
1336
+ "learning_rate": 2.3183377629730963e-06,
1337
+ "loss": 2.5435,
1338
+ "mean_token_accuracy": 0.507614666223526,
1339
+ "num_tokens": 41919.0,
1340
+ "step": 1330
1341
+ },
1342
+ {
1343
+ "entropy": 3.458651614189148,
1344
+ "epoch": 3.4536082474226806,
1345
+ "grad_norm": 57.5,
1346
+ "learning_rate": 2.28096952253398e-06,
1347
+ "loss": 3.842,
1348
+ "mean_token_accuracy": 0.39763966798782346,
1349
+ "num_tokens": 42237.0,
1350
+ "step": 1340
1351
+ },
1352
+ {
1353
+ "entropy": 3.4523784637451174,
1354
+ "epoch": 3.479381443298969,
1355
+ "grad_norm": 43.5,
1356
+ "learning_rate": 2.2436505345747505e-06,
1357
+ "loss": 3.1589,
1358
+ "mean_token_accuracy": 0.4656664371490479,
1359
+ "num_tokens": 42575.0,
1360
+ "step": 1350
1361
+ },
1362
+ {
1363
+ "entropy": 3.239696431159973,
1364
+ "epoch": 3.5051546391752577,
1365
+ "grad_norm": 74.0,
1366
+ "learning_rate": 2.2063891908622767e-06,
1367
+ "loss": 2.7964,
1368
+ "mean_token_accuracy": 0.48300274908542634,
1369
+ "num_tokens": 42936.0,
1370
+ "step": 1360
1371
+ },
1372
+ {
1373
+ "entropy": 3.3155257225036623,
1374
+ "epoch": 3.5309278350515463,
1375
+ "grad_norm": 58.75,
1376
+ "learning_rate": 2.169193870201203e-06,
1377
+ "loss": 3.106,
1378
+ "mean_token_accuracy": 0.463299959897995,
1379
+ "num_tokens": 43256.0,
1380
+ "step": 1370
1381
+ },
1382
+ {
1383
+ "entropy": 3.298970675468445,
1384
+ "epoch": 3.556701030927835,
1385
+ "grad_norm": 99.5,
1386
+ "learning_rate": 2.1320729365498404e-06,
1387
+ "loss": 3.3008,
1388
+ "mean_token_accuracy": 0.4579169362783432,
1389
+ "num_tokens": 43519.0,
1390
+ "step": 1380
1391
+ },
1392
+ {
1393
+ "entropy": 3.373097324371338,
1394
+ "epoch": 3.582474226804124,
1395
+ "grad_norm": 61.0,
1396
+ "learning_rate": 2.095034737139404e-06,
1397
+ "loss": 3.068,
1398
+ "mean_token_accuracy": 0.4262126713991165,
1399
+ "num_tokens": 43814.0,
1400
+ "step": 1390
1401
+ },
1402
+ {
1403
+ "entropy": 3.1966411590576174,
1404
+ "epoch": 3.6082474226804124,
1405
+ "grad_norm": 70.5,
1406
+ "learning_rate": 2.058087600596997e-06,
1407
+ "loss": 2.8911,
1408
+ "mean_token_accuracy": 0.5069397330284119,
1409
+ "num_tokens": 44153.0,
1410
+ "step": 1400
1411
+ },
1412
+ {
1413
+ "entropy": 3.3376646757125856,
1414
+ "epoch": 3.634020618556701,
1415
+ "grad_norm": 64.5,
1416
+ "learning_rate": 2.021239835072794e-06,
1417
+ "loss": 3.0766,
1418
+ "mean_token_accuracy": 0.46726988852024076,
1419
+ "num_tokens": 44465.0,
1420
+ "step": 1410
1421
+ },
1422
+ {
1423
+ "entropy": 3.3123555183410645,
1424
+ "epoch": 3.6597938144329896,
1425
+ "grad_norm": 53.5,
1426
+ "learning_rate": 1.984499726371819e-06,
1427
+ "loss": 2.8527,
1428
+ "mean_token_accuracy": 0.5056732088327408,
1429
+ "num_tokens": 44792.0,
1430
+ "step": 1420
1431
+ },
1432
+ {
1433
+ "entropy": 3.2575713872909544,
1434
+ "epoch": 3.6855670103092786,
1435
+ "grad_norm": 92.0,
1436
+ "learning_rate": 1.947875536090748e-06,
1437
+ "loss": 3.0459,
1438
+ "mean_token_accuracy": 0.46606201231479644,
1439
+ "num_tokens": 45093.0,
1440
+ "step": 1430
1441
+ },
1442
+ {
1443
+ "entropy": 3.2526177883148195,
1444
+ "epoch": 3.711340206185567,
1445
+ "grad_norm": 56.0,
1446
+ "learning_rate": 1.9113754997601614e-06,
1447
+ "loss": 3.2644,
1448
+ "mean_token_accuracy": 0.4528957188129425,
1449
+ "num_tokens": 45381.0,
1450
+ "step": 1440
1451
+ },
1452
+ {
1453
+ "entropy": 3.166100788116455,
1454
+ "epoch": 3.7371134020618557,
1455
+ "grad_norm": 60.75,
1456
+ "learning_rate": 1.875007824992654e-06,
1457
+ "loss": 2.9971,
1458
+ "mean_token_accuracy": 0.4395914673805237,
1459
+ "num_tokens": 45723.0,
1460
+ "step": 1450
1461
+ },
1462
+ {
1463
+ "entropy": 3.41998028755188,
1464
+ "epoch": 3.7628865979381443,
1465
+ "grad_norm": 75.0,
1466
+ "learning_rate": 1.8387806896372206e-06,
1467
+ "loss": 3.2834,
1468
+ "mean_token_accuracy": 0.4282302588224411,
1469
+ "num_tokens": 46053.0,
1470
+ "step": 1460
1471
+ },
1472
+ {
1473
+ "entropy": 3.263835144042969,
1474
+ "epoch": 3.788659793814433,
1475
+ "grad_norm": 73.0,
1476
+ "learning_rate": 1.8027022399403377e-06,
1477
+ "loss": 3.1976,
1478
+ "mean_token_accuracy": 0.4730199307203293,
1479
+ "num_tokens": 46330.0,
1480
+ "step": 1470
1481
+ },
1482
+ {
1483
+ "entropy": 3.38959059715271,
1484
+ "epoch": 3.8144329896907214,
1485
+ "grad_norm": 86.5,
1486
+ "learning_rate": 1.7667805887141526e-06,
1487
+ "loss": 3.072,
1488
+ "mean_token_accuracy": 0.48471441566944123,
1489
+ "num_tokens": 46675.0,
1490
+ "step": 1480
1491
+ },
1492
+ {
1493
+ "entropy": 3.516207385063171,
1494
+ "epoch": 3.8402061855670104,
1495
+ "grad_norm": 69.5,
1496
+ "learning_rate": 1.731023813512186e-06,
1497
+ "loss": 3.1848,
1498
+ "mean_token_accuracy": 0.4615781009197235,
1499
+ "num_tokens": 47005.0,
1500
+ "step": 1490
1501
+ },
1502
+ {
1503
+ "entropy": 3.435403323173523,
1504
+ "epoch": 3.865979381443299,
1505
+ "grad_norm": 62.5,
1506
+ "learning_rate": 1.695439954812968e-06,
1507
+ "loss": 3.5288,
1508
+ "mean_token_accuracy": 0.4129458874464035,
1509
+ "num_tokens": 47320.0,
1510
+ "step": 1500
1511
+ },
1512
+ {
1513
+ "entropy": 3.0035399436950683,
1514
+ "epoch": 3.8917525773195876,
1515
+ "grad_norm": 87.5,
1516
+ "learning_rate": 1.660037014212009e-06,
1517
+ "loss": 2.8988,
1518
+ "mean_token_accuracy": 0.5202886313199997,
1519
+ "num_tokens": 47654.0,
1520
+ "step": 1510
1521
+ },
1522
+ {
1523
+ "entropy": 3.282400608062744,
1524
+ "epoch": 3.917525773195876,
1525
+ "grad_norm": 68.0,
1526
+ "learning_rate": 1.62482295262251e-06,
1527
+ "loss": 3.0969,
1528
+ "mean_token_accuracy": 0.42603813409805297,
1529
+ "num_tokens": 47988.0,
1530
+ "step": 1520
1531
+ },
1532
+ {
1533
+ "entropy": 3.5067455053329466,
1534
+ "epoch": 3.943298969072165,
1535
+ "grad_norm": 63.75,
1536
+ "learning_rate": 1.589805688485231e-06,
1537
+ "loss": 3.2709,
1538
+ "mean_token_accuracy": 0.4377409517765045,
1539
+ "num_tokens": 48303.0,
1540
+ "step": 1530
1541
+ },
1542
+ {
1543
+ "entropy": 3.245515561103821,
1544
+ "epoch": 3.9690721649484537,
1545
+ "grad_norm": 55.0,
1546
+ "learning_rate": 1.5549930959878996e-06,
1547
+ "loss": 2.9502,
1548
+ "mean_token_accuracy": 0.5143730461597442,
1549
+ "num_tokens": 48637.0,
1550
+ "step": 1540
1551
+ },
1552
+ {
1553
+ "entropy": 3.1619110345840453,
1554
+ "epoch": 3.9948453608247423,
1555
+ "grad_norm": 61.25,
1556
+ "learning_rate": 1.5203930032945765e-06,
1557
+ "loss": 2.883,
1558
+ "mean_token_accuracy": 0.4930521368980408,
1559
+ "num_tokens": 48955.0,
1560
+ "step": 1550
1561
+ },
1562
+ {
1563
+ "entropy": 3.368650484085083,
1564
+ "epoch": 4.020618556701031,
1565
+ "grad_norm": 59.75,
1566
+ "learning_rate": 1.4860131907853664e-06,
1567
+ "loss": 3.0012,
1568
+ "mean_token_accuracy": 0.4367584019899368,
1569
+ "num_tokens": 49264.0,
1570
+ "step": 1560
1571
+ },
1572
+ {
1573
+ "entropy": 3.2799517631530763,
1574
+ "epoch": 4.046391752577319,
1575
+ "grad_norm": 107.0,
1576
+ "learning_rate": 1.45186138930688e-06,
1577
+ "loss": 3.0185,
1578
+ "mean_token_accuracy": 0.45549334287643434,
1579
+ "num_tokens": 49528.0,
1580
+ "step": 1570
1581
+ },
1582
+ {
1583
+ "entropy": 3.293484115600586,
1584
+ "epoch": 4.072164948453608,
1585
+ "grad_norm": 70.0,
1586
+ "learning_rate": 1.4179452784338265e-06,
1587
+ "loss": 3.3586,
1588
+ "mean_token_accuracy": 0.41464213728904725,
1589
+ "num_tokens": 49823.0,
1590
+ "step": 1580
1591
+ },
1592
+ {
1593
+ "entropy": 3.212877941131592,
1594
+ "epoch": 4.097938144329897,
1595
+ "grad_norm": 58.0,
1596
+ "learning_rate": 1.3842724847421435e-06,
1597
+ "loss": 2.9933,
1598
+ "mean_token_accuracy": 0.41281671822071075,
1599
+ "num_tokens": 50209.0,
1600
+ "step": 1590
1601
+ },
1602
+ {
1603
+ "entropy": 3.2077365159988402,
1604
+ "epoch": 4.123711340206185,
1605
+ "grad_norm": 60.0,
1606
+ "learning_rate": 1.3508505800940327e-06,
1607
+ "loss": 3.2048,
1608
+ "mean_token_accuracy": 0.45703212916851044,
1609
+ "num_tokens": 50566.0,
1610
+ "step": 1600
1611
+ },
1612
+ {
1613
+ "entropy": 3.2609221935272217,
1614
+ "epoch": 4.149484536082475,
1615
+ "grad_norm": 87.5,
1616
+ "learning_rate": 1.317687079935317e-06,
1617
+ "loss": 3.2903,
1618
+ "mean_token_accuracy": 0.45516538321971894,
1619
+ "num_tokens": 50932.0,
1620
+ "step": 1610
1621
+ },
1622
+ {
1623
+ "entropy": 3.199517750740051,
1624
+ "epoch": 4.175257731958763,
1625
+ "grad_norm": 121.5,
1626
+ "learning_rate": 1.2847894416054645e-06,
1627
+ "loss": 2.9641,
1628
+ "mean_token_accuracy": 0.49091013371944425,
1629
+ "num_tokens": 51242.0,
1630
+ "step": 1620
1631
+ },
1632
+ {
1633
+ "entropy": 3.2298731803894043,
1634
+ "epoch": 4.201030927835052,
1635
+ "grad_norm": 62.75,
1636
+ "learning_rate": 1.2521650626606926e-06,
1637
+ "loss": 2.9035,
1638
+ "mean_token_accuracy": 0.5309109538793564,
1639
+ "num_tokens": 51519.0,
1640
+ "step": 1630
1641
+ },
1642
+ {
1643
+ "entropy": 3.3670015573501586,
1644
+ "epoch": 4.22680412371134,
1645
+ "grad_norm": 80.0,
1646
+ "learning_rate": 1.219821279210507e-06,
1647
+ "loss": 3.0805,
1648
+ "mean_token_accuracy": 0.444789519906044,
1649
+ "num_tokens": 51823.0,
1650
+ "step": 1640
1651
+ },
1652
+ {
1653
+ "entropy": 3.3619128465652466,
1654
+ "epoch": 4.252577319587629,
1655
+ "grad_norm": 89.5,
1656
+ "learning_rate": 1.1877653642680618e-06,
1657
+ "loss": 3.0084,
1658
+ "mean_token_accuracy": 0.45184328258037565,
1659
+ "num_tokens": 52106.0,
1660
+ "step": 1650
1661
+ },
1662
+ {
1663
+ "entropy": 3.4395506143569947,
1664
+ "epoch": 4.278350515463917,
1665
+ "grad_norm": 72.0,
1666
+ "learning_rate": 1.1560045261147079e-06,
1667
+ "loss": 3.4761,
1668
+ "mean_token_accuracy": 0.41549868881702423,
1669
+ "num_tokens": 52494.0,
1670
+ "step": 1660
1671
+ },
1672
+ {
1673
+ "entropy": 3.154675102233887,
1674
+ "epoch": 4.304123711340206,
1675
+ "grad_norm": 62.25,
1676
+ "learning_rate": 1.1245459066790962e-06,
1677
+ "loss": 2.9656,
1678
+ "mean_token_accuracy": 0.49724196195602416,
1679
+ "num_tokens": 52842.0,
1680
+ "step": 1670
1681
+ },
1682
+ {
1683
+ "entropy": 3.107476019859314,
1684
+ "epoch": 4.329896907216495,
1685
+ "grad_norm": 87.5,
1686
+ "learning_rate": 1.0933965799312015e-06,
1687
+ "loss": 2.8218,
1688
+ "mean_token_accuracy": 0.5045855909585952,
1689
+ "num_tokens": 53174.0,
1690
+ "step": 1680
1691
+ },
1692
+ {
1693
+ "entropy": 3.290882110595703,
1694
+ "epoch": 4.355670103092783,
1695
+ "grad_norm": 87.0,
1696
+ "learning_rate": 1.062563550291626e-06,
1697
+ "loss": 3.0095,
1698
+ "mean_token_accuracy": 0.48295737206935885,
1699
+ "num_tokens": 53410.0,
1700
+ "step": 1690
1701
+ },
1702
+ {
1703
+ "entropy": 3.23433403968811,
1704
+ "epoch": 4.381443298969073,
1705
+ "grad_norm": 57.5,
1706
+ "learning_rate": 1.0320537510565474e-06,
1707
+ "loss": 3.0136,
1708
+ "mean_token_accuracy": 0.4336048990488052,
1709
+ "num_tokens": 53745.0,
1710
+ "step": 1700
1711
+ },
1712
+ {
1713
+ "entropy": 3.3636296510696413,
1714
+ "epoch": 4.407216494845361,
1715
+ "grad_norm": 60.5,
1716
+ "learning_rate": 1.0018740428386562e-06,
1717
+ "loss": 3.0679,
1718
+ "mean_token_accuracy": 0.47437537312507627,
1719
+ "num_tokens": 54031.0,
1720
+ "step": 1710
1721
+ },
1722
+ {
1723
+ "entropy": 3.443432426452637,
1724
+ "epoch": 4.43298969072165,
1725
+ "grad_norm": 86.5,
1726
+ "learning_rate": 9.720312120244368e-07,
1727
+ "loss": 3.2067,
1728
+ "mean_token_accuracy": 0.4474021762609482,
1729
+ "num_tokens": 54356.0,
1730
+ "step": 1720
1731
+ },
1732
+ {
1733
+ "entropy": 3.4575706720352173,
1734
+ "epoch": 4.458762886597938,
1735
+ "grad_norm": 80.0,
1736
+ "learning_rate": 9.425319692481421e-07,
1737
+ "loss": 3.244,
1738
+ "mean_token_accuracy": 0.44102594554424285,
1739
+ "num_tokens": 54627.0,
1740
+ "step": 1730
1741
+ },
1742
+ {
1743
+ "entropy": 3.0440264463424684,
1744
+ "epoch": 4.484536082474227,
1745
+ "grad_norm": 63.0,
1746
+ "learning_rate": 9.133829478828e-07,
1747
+ "loss": 2.9248,
1748
+ "mean_token_accuracy": 0.5230745673179626,
1749
+ "num_tokens": 54941.0,
1750
+ "step": 1740
1751
+ },
1752
+ {
1753
+ "entropy": 3.0223827600479125,
1754
+ "epoch": 4.510309278350515,
1755
+ "grad_norm": 77.5,
1756
+ "learning_rate": 8.845907025485945e-07,
1757
+ "loss": 2.8445,
1758
+ "mean_token_accuracy": 0.5213218927383423,
1759
+ "num_tokens": 55227.0,
1760
+ "step": 1750
1761
+ },
1762
+ {
1763
+ "entropy": 3.4792301177978517,
1764
+ "epoch": 4.536082474226804,
1765
+ "grad_norm": 63.0,
1766
+ "learning_rate": 8.561617076389556e-07,
1767
+ "loss": 3.1867,
1768
+ "mean_token_accuracy": 0.4527750164270401,
1769
+ "num_tokens": 55547.0,
1770
+ "step": 1760
1771
+ },
1772
+ {
1773
+ "entropy": 3.3560336589813233,
1774
+ "epoch": 4.561855670103093,
1775
+ "grad_norm": 85.0,
1776
+ "learning_rate": 8.281023558646892e-07,
1777
+ "loss": 3.2218,
1778
+ "mean_token_accuracy": 0.4219012975692749,
1779
+ "num_tokens": 55880.0,
1780
+ "step": 1770
1781
+ },
1782
+ {
1783
+ "entropy": 3.2443851470947265,
1784
+ "epoch": 4.587628865979381,
1785
+ "grad_norm": 72.0,
1786
+ "learning_rate": 8.004189568164721e-07,
1787
+ "loss": 3.2558,
1788
+ "mean_token_accuracy": 0.4505341827869415,
1789
+ "num_tokens": 56225.0,
1790
+ "step": 1780
1791
+ },
1792
+ {
1793
+ "entropy": 3.254337120056152,
1794
+ "epoch": 4.61340206185567,
1795
+ "grad_norm": 75.5,
1796
+ "learning_rate": 7.731177355460456e-07,
1797
+ "loss": 3.1863,
1798
+ "mean_token_accuracy": 0.4375568628311157,
1799
+ "num_tokens": 56528.0,
1800
+ "step": 1790
1801
+ },
1802
+ {
1803
+ "entropy": 3.0034953951835632,
1804
+ "epoch": 4.639175257731958,
1805
+ "grad_norm": 64.0,
1806
+ "learning_rate": 7.462048311664086e-07,
1807
+ "loss": 2.6958,
1808
+ "mean_token_accuracy": 0.536104878783226,
1809
+ "num_tokens": 56847.0,
1810
+ "step": 1800
1811
+ },
1812
+ {
1813
+ "entropy": 3.2068035364151,
1814
+ "epoch": 4.664948453608248,
1815
+ "grad_norm": 85.0,
1816
+ "learning_rate": 7.196862954713438e-07,
1817
+ "loss": 2.9987,
1818
+ "mean_token_accuracy": 0.4882556527853012,
1819
+ "num_tokens": 57154.0,
1820
+ "step": 1810
1821
+ },
1822
+ {
1823
+ "entropy": 3.0603734254837036,
1824
+ "epoch": 4.690721649484536,
1825
+ "grad_norm": 54.5,
1826
+ "learning_rate": 6.935680915745743e-07,
1827
+ "loss": 2.6614,
1828
+ "mean_token_accuracy": 0.4894832164049149,
1829
+ "num_tokens": 57445.0,
1830
+ "step": 1820
1831
+ },
1832
+ {
1833
+ "entropy": 3.4436551094055177,
1834
+ "epoch": 4.716494845360825,
1835
+ "grad_norm": 66.0,
1836
+ "learning_rate": 6.678560925688629e-07,
1837
+ "loss": 3.4491,
1838
+ "mean_token_accuracy": 0.42854584753513336,
1839
+ "num_tokens": 57799.0,
1840
+ "step": 1830
1841
+ },
1842
+ {
1843
+ "entropy": 3.3284261226654053,
1844
+ "epoch": 4.742268041237113,
1845
+ "grad_norm": 69.5,
1846
+ "learning_rate": 6.425560802053551e-07,
1847
+ "loss": 3.1212,
1848
+ "mean_token_accuracy": 0.4246213287115097,
1849
+ "num_tokens": 58153.0,
1850
+ "step": 1840
1851
+ },
1852
+ {
1853
+ "entropy": 3.3664702415466308,
1854
+ "epoch": 4.768041237113402,
1855
+ "grad_norm": 85.5,
1856
+ "learning_rate": 6.176737435934593e-07,
1857
+ "loss": 2.97,
1858
+ "mean_token_accuracy": 0.45907922089099884,
1859
+ "num_tokens": 58473.0,
1860
+ "step": 1850
1861
+ },
1862
+ {
1863
+ "entropy": 3.260459637641907,
1864
+ "epoch": 4.793814432989691,
1865
+ "grad_norm": 70.0,
1866
+ "learning_rate": 5.932146779215614e-07,
1867
+ "loss": 3.091,
1868
+ "mean_token_accuracy": 0.47388018369674684,
1869
+ "num_tokens": 58787.0,
1870
+ "step": 1860
1871
+ },
1872
+ {
1873
+ "entropy": 3.260399317741394,
1874
+ "epoch": 4.819587628865979,
1875
+ "grad_norm": 64.0,
1876
+ "learning_rate": 5.691843831988547e-07,
1877
+ "loss": 3.1014,
1878
+ "mean_token_accuracy": 0.4497444421052933,
1879
+ "num_tokens": 59126.0,
1880
+ "step": 1870
1881
+ },
1882
+ {
1883
+ "entropy": 3.1505658864974975,
1884
+ "epoch": 4.845360824742268,
1885
+ "grad_norm": 101.5,
1886
+ "learning_rate": 5.45588263018581e-07,
1887
+ "loss": 3.0726,
1888
+ "mean_token_accuracy": 0.4976365238428116,
1889
+ "num_tokens": 59480.0,
1890
+ "step": 1880
1891
+ },
1892
+ {
1893
+ "entropy": 3.3082144498825072,
1894
+ "epoch": 4.871134020618557,
1895
+ "grad_norm": 55.25,
1896
+ "learning_rate": 5.224316233429422e-07,
1897
+ "loss": 3.4489,
1898
+ "mean_token_accuracy": 0.41699815094470977,
1899
+ "num_tokens": 59755.0,
1900
+ "step": 1890
1901
+ },
1902
+ {
1903
+ "entropy": 3.4492565393447876,
1904
+ "epoch": 4.896907216494846,
1905
+ "grad_norm": 129.0,
1906
+ "learning_rate": 4.997196713099728e-07,
1907
+ "loss": 3.3794,
1908
+ "mean_token_accuracy": 0.44576013684272764,
1909
+ "num_tokens": 60046.0,
1910
+ "step": 1900
1911
+ },
1912
+ {
1913
+ "entropy": 3.398744559288025,
1914
+ "epoch": 4.922680412371134,
1915
+ "grad_norm": 88.0,
1916
+ "learning_rate": 4.774575140626317e-07,
1917
+ "loss": 3.3877,
1918
+ "mean_token_accuracy": 0.4413867056369781,
1919
+ "num_tokens": 60384.0,
1920
+ "step": 1910
1921
+ },
1922
+ {
1923
+ "entropy": 3.20350604057312,
1924
+ "epoch": 4.948453608247423,
1925
+ "grad_norm": 56.0,
1926
+ "learning_rate": 4.556501576003791e-07,
1927
+ "loss": 3.0359,
1928
+ "mean_token_accuracy": 0.45435314774513247,
1929
+ "num_tokens": 60655.0,
1930
+ "step": 1920
1931
+ },
1932
+ {
1933
+ "entropy": 2.980875849723816,
1934
+ "epoch": 4.974226804123711,
1935
+ "grad_norm": 80.5,
1936
+ "learning_rate": 4.343025056534994e-07,
1937
+ "loss": 2.9326,
1938
+ "mean_token_accuracy": 0.5152811527252197,
1939
+ "num_tokens": 60949.0,
1940
+ "step": 1930
1941
+ },
1942
+ {
1943
+ "entropy": 3.3593399286270142,
1944
+ "epoch": 5.0,
1945
+ "grad_norm": 70.5,
1946
+ "learning_rate": 4.134193585804197e-07,
1947
+ "loss": 3.2147,
1948
+ "mean_token_accuracy": 0.4529120117425919,
1949
+ "num_tokens": 61260.0,
1950
+ "step": 1940
1951
+ },
1952
+ {
1953
+ "entropy": 3.3000378370285035,
1954
+ "epoch": 5.025773195876289,
1955
+ "grad_norm": 75.0,
1956
+ "learning_rate": 3.930054122882709e-07,
1957
+ "loss": 3.2406,
1958
+ "mean_token_accuracy": 0.4164623826742172,
1959
+ "num_tokens": 61558.0,
1960
+ "step": 1950
1961
+ },
1962
+ {
1963
+ "entropy": 3.211987781524658,
1964
+ "epoch": 5.051546391752577,
1965
+ "grad_norm": 56.75,
1966
+ "learning_rate": 3.7306525717694195e-07,
1967
+ "loss": 3.1386,
1968
+ "mean_token_accuracy": 0.5111187756061554,
1969
+ "num_tokens": 61848.0,
1970
+ "step": 1960
1971
+ },
1972
+ {
1973
+ "entropy": 3.2421340942382812,
1974
+ "epoch": 5.077319587628866,
1975
+ "grad_norm": 79.5,
1976
+ "learning_rate": 3.536033771068506e-07,
1977
+ "loss": 2.9573,
1978
+ "mean_token_accuracy": 0.49586873054504393,
1979
+ "num_tokens": 62126.0,
1980
+ "step": 1970
1981
+ },
1982
+ {
1983
+ "entropy": 3.501763868331909,
1984
+ "epoch": 5.103092783505154,
1985
+ "grad_norm": 67.5,
1986
+ "learning_rate": 3.3462414839068233e-07,
1987
+ "loss": 3.2128,
1988
+ "mean_token_accuracy": 0.4240656703710556,
1989
+ "num_tokens": 62529.0,
1990
+ "step": 1980
1991
+ },
1992
+ {
1993
+ "entropy": 3.2703943729400633,
1994
+ "epoch": 5.128865979381443,
1995
+ "grad_norm": 57.5,
1996
+ "learning_rate": 3.1613183880930124e-07,
1997
+ "loss": 3.1511,
1998
+ "mean_token_accuracy": 0.4509860217571259,
1999
+ "num_tokens": 62808.0,
2000
+ "step": 1990
2001
+ },
2002
+ {
2003
+ "entropy": 3.1918478488922117,
2004
+ "epoch": 5.154639175257732,
2005
+ "grad_norm": 90.0,
2006
+ "learning_rate": 2.9813060665207615e-07,
2007
+ "loss": 2.9956,
2008
+ "mean_token_accuracy": 0.5052088230848313,
2009
+ "num_tokens": 63054.0,
2010
+ "step": 2000
2011
+ },
2012
+ {
2013
+ "entropy": 3.2867510318756104,
2014
+ "epoch": 5.180412371134021,
2015
+ "grad_norm": 62.0,
2016
+ "learning_rate": 2.8062449978182056e-07,
2017
+ "loss": 3.2784,
2018
+ "mean_token_accuracy": 0.4433626294136047,
2019
+ "num_tokens": 63378.0,
2020
+ "step": 2010
2021
+ },
2022
+ {
2023
+ "entropy": 3.276996612548828,
2024
+ "epoch": 5.206185567010309,
2025
+ "grad_norm": 59.5,
2026
+ "learning_rate": 2.636174547245671e-07,
2027
+ "loss": 2.9721,
2028
+ "mean_token_accuracy": 0.4633091241121292,
2029
+ "num_tokens": 63698.0,
2030
+ "step": 2020
2031
+ },
2032
+ {
2033
+ "entropy": 3.3924339771270753,
2034
+ "epoch": 5.231958762886598,
2035
+ "grad_norm": 70.5,
2036
+ "learning_rate": 2.471132957843775e-07,
2037
+ "loss": 3.5247,
2038
+ "mean_token_accuracy": 0.40621828436851504,
2039
+ "num_tokens": 64026.0,
2040
+ "step": 2030
2041
+ },
2042
+ {
2043
+ "entropy": 3.183038115501404,
2044
+ "epoch": 5.257731958762887,
2045
+ "grad_norm": 64.0,
2046
+ "learning_rate": 2.3111573418338724e-07,
2047
+ "loss": 3.057,
2048
+ "mean_token_accuracy": 0.4755241394042969,
2049
+ "num_tokens": 64357.0,
2050
+ "step": 2040
2051
+ },
2052
+ {
2053
+ "entropy": 3.0321537017822267,
2054
+ "epoch": 5.283505154639175,
2055
+ "grad_norm": 80.0,
2056
+ "learning_rate": 2.156283672272777e-07,
2057
+ "loss": 3.0536,
2058
+ "mean_token_accuracy": 0.46814982295036317,
2059
+ "num_tokens": 64717.0,
2060
+ "step": 2050
2061
+ },
2062
+ {
2063
+ "entropy": 3.307017612457275,
2064
+ "epoch": 5.309278350515464,
2065
+ "grad_norm": 50.5,
2066
+ "learning_rate": 2.0065467749636497e-07,
2067
+ "loss": 3.0379,
2068
+ "mean_token_accuracy": 0.4797358334064484,
2069
+ "num_tokens": 65004.0,
2070
+ "step": 2060
2071
+ },
2072
+ {
2073
+ "entropy": 3.342577505111694,
2074
+ "epoch": 5.335051546391752,
2075
+ "grad_norm": 103.5,
2076
+ "learning_rate": 1.861980320624873e-07,
2077
+ "loss": 3.111,
2078
+ "mean_token_accuracy": 0.46001616716384885,
2079
+ "num_tokens": 65327.0,
2080
+ "step": 2070
2081
+ },
2082
+ {
2083
+ "entropy": 3.2767983198165895,
2084
+ "epoch": 5.360824742268041,
2085
+ "grad_norm": 51.25,
2086
+ "learning_rate": 1.7226168173186396e-07,
2087
+ "loss": 3.2626,
2088
+ "mean_token_accuracy": 0.44220549464225767,
2089
+ "num_tokens": 65743.0,
2090
+ "step": 2080
2091
+ },
2092
+ {
2093
+ "entropy": 3.1632169485092163,
2094
+ "epoch": 5.38659793814433,
2095
+ "grad_norm": 80.0,
2096
+ "learning_rate": 1.5884876031410202e-07,
2097
+ "loss": 3.4269,
2098
+ "mean_token_accuracy": 0.43744454979896547,
2099
+ "num_tokens": 66042.0,
2100
+ "step": 2090
2101
+ },
2102
+ {
2103
+ "entropy": 3.083737111091614,
2104
+ "epoch": 5.412371134020619,
2105
+ "grad_norm": 76.0,
2106
+ "learning_rate": 1.4596228391750676e-07,
2107
+ "loss": 3.0883,
2108
+ "mean_token_accuracy": 0.4829647660255432,
2109
+ "num_tokens": 66327.0,
2110
+ "step": 2100
2111
+ },
2112
+ {
2113
+ "entropy": 3.032327151298523,
2114
+ "epoch": 5.438144329896907,
2115
+ "grad_norm": 99.5,
2116
+ "learning_rate": 1.3360515027086462e-07,
2117
+ "loss": 2.8775,
2118
+ "mean_token_accuracy": 0.4383985221385956,
2119
+ "num_tokens": 66641.0,
2120
+ "step": 2110
2121
+ },
2122
+ {
2123
+ "entropy": 3.05249879360199,
2124
+ "epoch": 5.463917525773196,
2125
+ "grad_norm": 87.0,
2126
+ "learning_rate": 1.2178013807184085e-07,
2127
+ "loss": 3.0836,
2128
+ "mean_token_accuracy": 0.457595694065094,
2129
+ "num_tokens": 66937.0,
2130
+ "step": 2120
2131
+ },
2132
+ {
2133
+ "entropy": 3.3152157068252563,
2134
+ "epoch": 5.489690721649485,
2135
+ "grad_norm": 76.0,
2136
+ "learning_rate": 1.1048990636214618e-07,
2137
+ "loss": 3.1917,
2138
+ "mean_token_accuracy": 0.48176924884319305,
2139
+ "num_tokens": 67223.0,
2140
+ "step": 2130
2141
+ },
2142
+ {
2143
+ "entropy": 3.5079283475875855,
2144
+ "epoch": 5.515463917525773,
2145
+ "grad_norm": 60.75,
2146
+ "learning_rate": 9.973699392960917e-08,
2147
+ "loss": 3.3596,
2148
+ "mean_token_accuracy": 0.40318597555160524,
2149
+ "num_tokens": 67536.0,
2150
+ "step": 2140
2151
+ },
2152
+ {
2153
+ "entropy": 3.342143940925598,
2154
+ "epoch": 5.541237113402062,
2155
+ "grad_norm": 65.5,
2156
+ "learning_rate": 8.952381873728821e-08,
2157
+ "loss": 2.9796,
2158
+ "mean_token_accuracy": 0.46548409163951876,
2159
+ "num_tokens": 67876.0,
2160
+ "step": 2150
2161
+ },
2162
+ {
2163
+ "entropy": 3.1259532451629637,
2164
+ "epoch": 5.56701030927835,
2165
+ "grad_norm": 49.75,
2166
+ "learning_rate": 7.985267737975588e-08,
2167
+ "loss": 3.0859,
2168
+ "mean_token_accuracy": 0.4911260426044464,
2169
+ "num_tokens": 68229.0,
2170
+ "step": 2160
2171
+ },
2172
+ {
2173
+ "entropy": 3.450204849243164,
2174
+ "epoch": 5.592783505154639,
2175
+ "grad_norm": 71.5,
2176
+ "learning_rate": 7.072574456667258e-08,
2177
+ "loss": 3.0189,
2178
+ "mean_token_accuracy": 0.47769983410835265,
2179
+ "num_tokens": 68567.0,
2180
+ "step": 2170
2181
+ },
2182
+ {
2183
+ "entropy": 3.2766273736953737,
2184
+ "epoch": 5.618556701030927,
2185
+ "grad_norm": 83.5,
2186
+ "learning_rate": 6.214507263376801e-08,
2187
+ "loss": 2.881,
2188
+ "mean_token_accuracy": 0.48972640335559847,
2189
+ "num_tokens": 68846.0,
2190
+ "step": 2180
2191
+ },
2192
+ {
2193
+ "entropy": 3.267438220977783,
2194
+ "epoch": 5.644329896907217,
2195
+ "grad_norm": 68.0,
2196
+ "learning_rate": 5.411259108134115e-08,
2197
+ "loss": 2.9084,
2198
+ "mean_token_accuracy": 0.4877107352018356,
2199
+ "num_tokens": 69197.0,
2200
+ "step": 2190
2201
+ },
2202
+ {
2203
+ "entropy": 3.320011782646179,
2204
+ "epoch": 5.670103092783505,
2205
+ "grad_norm": 70.0,
2206
+ "learning_rate": 4.663010614038205e-08,
2207
+ "loss": 3.1586,
2208
+ "mean_token_accuracy": 0.4482086032629013,
2209
+ "num_tokens": 69493.0,
2210
+ "step": 2200
2211
+ },
2212
+ {
2213
+ "entropy": 3.4605057001113892,
2214
+ "epoch": 5.695876288659794,
2215
+ "grad_norm": 86.0,
2216
+ "learning_rate": 3.9699300366410895e-08,
2217
+ "loss": 3.2727,
2218
+ "mean_token_accuracy": 0.4599607527256012,
2219
+ "num_tokens": 69832.0,
2220
+ "step": 2210
2221
+ },
2222
+ {
2223
+ "entropy": 3.275946855545044,
2224
+ "epoch": 5.721649484536083,
2225
+ "grad_norm": 58.5,
2226
+ "learning_rate": 3.332173226113067e-08,
2227
+ "loss": 3.0561,
2228
+ "mean_token_accuracy": 0.460858029127121,
2229
+ "num_tokens": 70215.0,
2230
+ "step": 2220
2231
+ },
2232
+ {
2233
+ "entropy": 3.389675521850586,
2234
+ "epoch": 5.747422680412371,
2235
+ "grad_norm": 92.0,
2236
+ "learning_rate": 2.7498835921971058e-08,
2237
+ "loss": 3.0778,
2238
+ "mean_token_accuracy": 0.44814312756061553,
2239
+ "num_tokens": 70534.0,
2240
+ "step": 2230
2241
+ },
2242
+ {
2243
+ "entropy": 3.1975868225097654,
2244
+ "epoch": 5.77319587628866,
2245
+ "grad_norm": 72.5,
2246
+ "learning_rate": 2.2231920719610057e-08,
2247
+ "loss": 2.9487,
2248
+ "mean_token_accuracy": 0.47108798921108247,
2249
+ "num_tokens": 70838.0,
2250
+ "step": 2240
2251
+ },
2252
+ {
2253
+ "entropy": 3.20022566318512,
2254
+ "epoch": 5.798969072164948,
2255
+ "grad_norm": 69.0,
2256
+ "learning_rate": 1.75221710035417e-08,
2257
+ "loss": 2.67,
2258
+ "mean_token_accuracy": 0.5278927952051162,
2259
+ "num_tokens": 71133.0,
2260
+ "step": 2250
2261
+ },
2262
+ {
2263
+ "entropy": 3.337849473953247,
2264
+ "epoch": 5.824742268041237,
2265
+ "grad_norm": 93.5,
2266
+ "learning_rate": 1.3370645835754926e-08,
2267
+ "loss": 3.1954,
2268
+ "mean_token_accuracy": 0.4186277240514755,
2269
+ "num_tokens": 71475.0,
2270
+ "step": 2260
2271
+ },
2272
+ {
2273
+ "entropy": 3.2620397567749024,
2274
+ "epoch": 5.850515463917525,
2275
+ "grad_norm": 99.0,
2276
+ "learning_rate": 9.77827875258769e-09,
2277
+ "loss": 3.0062,
2278
+ "mean_token_accuracy": 0.4634517639875412,
2279
+ "num_tokens": 71766.0,
2280
+ "step": 2270
2281
+ },
2282
+ {
2283
+ "entropy": 3.139668273925781,
2284
+ "epoch": 5.876288659793815,
2285
+ "grad_norm": 84.0,
2286
+ "learning_rate": 6.745877554806268e-09,
2287
+ "loss": 3.0198,
2288
+ "mean_token_accuracy": 0.4661685138940811,
2289
+ "num_tokens": 72053.0,
2290
+ "step": 2280
2291
+ },
2292
+ {
2293
+ "entropy": 3.293707489967346,
2294
+ "epoch": 5.902061855670103,
2295
+ "grad_norm": 59.0,
2296
+ "learning_rate": 4.274124125958068e-09,
2297
+ "loss": 3.0505,
2298
+ "mean_token_accuracy": 0.46237342357635497,
2299
+ "num_tokens": 72347.0,
2300
+ "step": 2290
2301
+ },
2302
+ {
2303
+ "entropy": 3.4691567182540894,
2304
+ "epoch": 5.927835051546392,
2305
+ "grad_norm": 59.5,
2306
+ "learning_rate": 2.363574279040104e-09,
2307
+ "loss": 3.3964,
2308
+ "mean_token_accuracy": 0.4291213423013687,
2309
+ "num_tokens": 72681.0,
2310
+ "step": 2300
2311
+ },
2312
+ {
2313
+ "entropy": 3.170962166786194,
2314
+ "epoch": 5.953608247422681,
2315
+ "grad_norm": 81.5,
2316
+ "learning_rate": 1.0146576315145374e-09,
2317
+ "loss": 2.9427,
2318
+ "mean_token_accuracy": 0.5170234173536301,
2319
+ "num_tokens": 72977.0,
2320
+ "step": 2310
2321
+ },
2322
+ {
2323
+ "entropy": 3.2913349628448487,
2324
+ "epoch": 5.979381443298969,
2325
+ "grad_norm": 67.0,
2326
+ "learning_rate": 2.276775087031724e-10,
2327
+ "loss": 3.0982,
2328
+ "mean_token_accuracy": 0.40257313251495364,
2329
+ "num_tokens": 73292.0,
2330
+ "step": 2320
2331
+ }
2332
+ ],
2333
+ "logging_steps": 10,
2334
+ "max_steps": 2328,
2335
+ "num_input_tokens_seen": 0,
2336
+ "num_train_epochs": 6,
2337
+ "save_steps": 100,
2338
+ "stateful_callbacks": {
2339
+ "TrainerControl": {
2340
+ "args": {
2341
+ "should_epoch_stop": false,
2342
+ "should_evaluate": false,
2343
+ "should_log": false,
2344
+ "should_save": true,
2345
+ "should_training_stop": true
2346
+ },
2347
+ "attributes": {}
2348
+ }
2349
+ },
2350
+ "total_flos": 621651308691456.0,
2351
+ "train_batch_size": 1,
2352
+ "trial_name": null,
2353
+ "trial_params": null
2354
+ }
checkpoint-2328/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:609c0100e8e8ae1a8654246f76d53314d98a0e4dcf6191e136010cdaddbb8430
3
+ size 6225
checkpoint-2328/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "dtype": "bfloat16",
8
+ "eos_token_id": 151643,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 2048,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 6144,
14
+ "layer_types": [
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention"
43
+ ],
44
+ "max_position_embeddings": 32768,
45
+ "max_window_layers": 28,
46
+ "model_type": "qwen3",
47
+ "num_attention_heads": 16,
48
+ "num_hidden_layers": 28,
49
+ "num_key_value_heads": 8,
50
+ "pad_token_id": 151643,
51
+ "rms_norm_eps": 1e-06,
52
+ "rope_scaling": null,
53
+ "rope_theta": 1000000,
54
+ "sliding_window": null,
55
+ "tie_word_embeddings": true,
56
+ "transformers_version": "4.57.3",
57
+ "use_cache": true,
58
+ "use_sliding_window": false,
59
+ "vocab_size": 151936
60
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token_id": [
3
+ 151643
4
+ ],
5
+ "max_new_tokens": 2048,
6
+ "pad_token_id": 151643,
7
+ "transformers_version": "4.57.3"
8
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:899a58d2b285510f699ed77c114a40c8e6b949d639706ecc55e1074ca1ae1d00
3
+ size 3441185608
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
tokenizer_config.json ADDED
@@ -0,0 +1,239 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|endoftext|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 131072,
235
+ "pad_token": "<|endoftext|>",
236
+ "split_special_tokens": false,
237
+ "tokenizer_class": "Qwen2Tokenizer",
238
+ "unk_token": null
239
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:609c0100e8e8ae1a8654246f76d53314d98a0e4dcf6191e136010cdaddbb8430
3
+ size 6225
vocab.json ADDED
The diff for this file is too large to render. See raw diff