Chiung-Yi commited on
Commit
a8aacfb
·
verified ·
1 Parent(s): 71863e5

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. README.md +149 -0
  3. adapter_config.json +42 -0
  4. adapter_model.safetensors +3 -0
  5. added_tokens.json +28 -0
  6. chat_template.jinja +86 -0
  7. checkpoint-105/README.md +208 -0
  8. checkpoint-105/adapter_config.json +42 -0
  9. checkpoint-105/adapter_model.safetensors +3 -0
  10. checkpoint-105/added_tokens.json +28 -0
  11. checkpoint-105/chat_template.jinja +86 -0
  12. checkpoint-105/merges.txt +0 -0
  13. checkpoint-105/optimizer.pt +3 -0
  14. checkpoint-105/rng_state_0.pth +3 -0
  15. checkpoint-105/rng_state_1.pth +3 -0
  16. checkpoint-105/scheduler.pt +3 -0
  17. checkpoint-105/special_tokens_map.json +31 -0
  18. checkpoint-105/tokenizer.json +3 -0
  19. checkpoint-105/tokenizer_config.json +239 -0
  20. checkpoint-105/trainer_state.json +1084 -0
  21. checkpoint-105/training_args.bin +3 -0
  22. checkpoint-105/vocab.json +0 -0
  23. checkpoint-210/README.md +208 -0
  24. checkpoint-210/adapter_config.json +42 -0
  25. checkpoint-210/adapter_model.safetensors +3 -0
  26. checkpoint-210/added_tokens.json +28 -0
  27. checkpoint-210/chat_template.jinja +86 -0
  28. checkpoint-210/merges.txt +0 -0
  29. checkpoint-210/optimizer.pt +3 -0
  30. checkpoint-210/rng_state_0.pth +3 -0
  31. checkpoint-210/rng_state_1.pth +3 -0
  32. checkpoint-210/scheduler.pt +3 -0
  33. checkpoint-210/special_tokens_map.json +31 -0
  34. checkpoint-210/tokenizer.json +3 -0
  35. checkpoint-210/tokenizer_config.json +239 -0
  36. checkpoint-210/trainer_state.json +2134 -0
  37. checkpoint-210/training_args.bin +3 -0
  38. checkpoint-210/vocab.json +0 -0
  39. checkpoint-315/README.md +208 -0
  40. checkpoint-315/adapter_config.json +42 -0
  41. checkpoint-315/adapter_model.safetensors +3 -0
  42. checkpoint-315/added_tokens.json +28 -0
  43. checkpoint-315/chat_template.jinja +86 -0
  44. checkpoint-315/merges.txt +0 -0
  45. checkpoint-315/optimizer.pt +3 -0
  46. checkpoint-315/rng_state_0.pth +3 -0
  47. checkpoint-315/rng_state_1.pth +3 -0
  48. checkpoint-315/scheduler.pt +3 -0
  49. checkpoint-315/special_tokens_map.json +31 -0
  50. checkpoint-315/tokenizer.json +3 -0
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint-105/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ checkpoint-210/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ checkpoint-315/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,149 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: apache-2.0
4
+ base_model: Qwen/Qwen3-4B-Instruct-2507
5
+ tags:
6
+ - axolotl
7
+ - base_model:adapter:Qwen/Qwen3-4B-Instruct-2507
8
+ - lora
9
+ - transformers
10
+ datasets:
11
+ - LuxMuseAI/StreetMath
12
+ pipeline_tag: text-generation
13
+ model-index:
14
+ - name: outputs/mymodel
15
+ results: []
16
+ ---
17
+
18
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
19
+ should probably proofread and complete it, then remove this comment. -->
20
+
21
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
22
+ <details><summary>See axolotl config</summary>
23
+
24
+ axolotl version: `0.13.0.dev0`
25
+ ```yaml
26
+ adapter: lora
27
+ base_model: Qwen/Qwen3-4B-Instruct-2507
28
+ bf16: auto
29
+ dataset_processes: 32
30
+ datasets:
31
+ - type: alpaca
32
+ message_property_mappings:
33
+ instruction: instruction
34
+ input: input
35
+ output: output
36
+ path: LuxMuseAI/StreetMath
37
+ trust_remote_code: false
38
+ gradient_accumulation_steps: 1
39
+ gradient_checkpointing: false
40
+ learning_rate: 0.0002
41
+ lisa_layers_attribute: model.layers
42
+ load_best_model_at_end: false
43
+ load_in_4bit: false
44
+ load_in_8bit: true
45
+ lora_alpha: 16
46
+ lora_dropout: 0.05
47
+ lora_r: 8
48
+ lora_target_modules:
49
+ - q_proj
50
+ - v_proj
51
+ - k_proj
52
+ - o_proj
53
+ - gate_proj
54
+ - down_proj
55
+ - up_proj
56
+ loraplus_lr_embedding: 1.0e-06
57
+ lr_scheduler: cosine
58
+ max_prompt_len: 512
59
+ mean_resizing_embeddings: false
60
+ micro_batch_size: 16
61
+ num_epochs: 3.0
62
+ optimizer: adamw_bnb_8bit
63
+ output_dir: ./outputs/mymodel
64
+ pretrain_multipack_attn: true
65
+ pretrain_multipack_buffer_size: 10000
66
+ qlora_sharded_model_loading: false
67
+ ray_num_workers: 1
68
+ resources_per_worker:
69
+ GPU: 1
70
+ sample_packing_bin_size: 200
71
+ sample_packing_group_size: 100000
72
+ save_only_model: false
73
+ save_safetensors: true
74
+ sequence_len: 4096
75
+ shuffle_merged_datasets: true
76
+ skip_prepare_dataset: false
77
+ strict: false
78
+ train_on_inputs: false
79
+ trl:
80
+ log_completions: false
81
+ ref_model_mixup_alpha: 0.9
82
+ ref_model_sync_steps: 64
83
+ sync_ref_model: false
84
+ use_vllm: false
85
+ vllm_device: auto
86
+ vllm_dtype: auto
87
+ vllm_gpu_memory_utilization: 0.9
88
+ use_ray: false
89
+ val_set_size: 0.0
90
+ weight_decay: 0.0
91
+
92
+ # Weights & Biases logging (optional)
93
+ wandb_project: StreetMath
94
+ wandb_entity: ctseng- # Replace with your actual entity name
95
+ wandb_watch: all
96
+ wandb_name: StreetMath-4B-Instruct-2507
97
+ wandb_log_model: "end"
98
+
99
+
100
+ ```
101
+
102
+ </details><br>
103
+
104
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/ctseng-/StreetMath/runs/2tnvvjjg)
105
+ # outputs/mymodel
106
+
107
+ This model is a fine-tuned version of [Qwen/Qwen3-4B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507) on the LuxMuseAI/StreetMath dataset.
108
+
109
+ ## Model description
110
+
111
+ More information needed
112
+
113
+ ## Intended uses & limitations
114
+
115
+ More information needed
116
+
117
+ ## Training and evaluation data
118
+
119
+ More information needed
120
+
121
+ ## Training procedure
122
+
123
+ ### Training hyperparameters
124
+
125
+ The following hyperparameters were used during training:
126
+ - learning_rate: 0.0002
127
+ - train_batch_size: 16
128
+ - eval_batch_size: 16
129
+ - seed: 42
130
+ - distributed_type: multi-GPU
131
+ - num_devices: 2
132
+ - total_train_batch_size: 32
133
+ - total_eval_batch_size: 32
134
+ - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
135
+ - lr_scheduler_type: cosine
136
+ - lr_scheduler_warmup_steps: 9
137
+ - training_steps: 315
138
+
139
+ ### Training results
140
+
141
+
142
+
143
+ ### Framework versions
144
+
145
+ - PEFT 0.17.0
146
+ - Transformers 4.55.2
147
+ - Pytorch 2.6.0+cu124
148
+ - Datasets 4.0.0
149
+ - Tokenizers 0.21.4
adapter_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen3-4B-Instruct-2507",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": null,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 8,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "q_proj",
29
+ "gate_proj",
30
+ "up_proj",
31
+ "v_proj",
32
+ "down_proj",
33
+ "k_proj",
34
+ "o_proj"
35
+ ],
36
+ "target_parameters": [],
37
+ "task_type": "CAUSAL_LM",
38
+ "trainable_token_indices": null,
39
+ "use_dora": false,
40
+ "use_qalora": false,
41
+ "use_rslora": false
42
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e45b31c137803c2351f0144a9d9cdee51846c1e3ed927ec07b20c1cc38ea3b8
3
+ size 66126768
added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
chat_template.jinja ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if message.content is string %}
27
+ {%- set content = message.content %}
28
+ {%- else %}
29
+ {%- set content = '' %}
30
+ {%- endif %}
31
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
+ {%- elif message.role == "assistant" %}
34
+ {%- set reasoning_content = '' %}
35
+ {%- if message.reasoning_content is string %}
36
+ {%- set reasoning_content = message.reasoning_content %}
37
+ {%- else %}
38
+ {%- if '</think>' in content %}
39
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
+ {%- endif %}
42
+ {%- endif %}
43
+ {%- if loop.index0 > ns.last_query_index %}
44
+ {%- if loop.last or (not loop.last and reasoning_content) %}
45
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
+ {%- else %}
47
+ {{- '<|im_start|>' + message.role + '\n' + content }}
48
+ {%- endif %}
49
+ {%- else %}
50
+ {{- '<|im_start|>' + message.role + '\n' + content }}
51
+ {%- endif %}
52
+ {%- if message.tool_calls %}
53
+ {%- for tool_call in message.tool_calls %}
54
+ {%- if (loop.first and content) or (not loop.first) %}
55
+ {{- '\n' }}
56
+ {%- endif %}
57
+ {%- if tool_call.function %}
58
+ {%- set tool_call = tool_call.function %}
59
+ {%- endif %}
60
+ {{- '<tool_call>\n{"name": "' }}
61
+ {{- tool_call.name }}
62
+ {{- '", "arguments": ' }}
63
+ {%- if tool_call.arguments is string %}
64
+ {{- tool_call.arguments }}
65
+ {%- else %}
66
+ {{- tool_call.arguments | tojson }}
67
+ {%- endif %}
68
+ {{- '}\n</tool_call>' }}
69
+ {%- endfor %}
70
+ {%- endif %}
71
+ {{- '<|im_end|>\n' }}
72
+ {%- elif message.role == "tool" %}
73
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
74
+ {{- '<|im_start|>user' }}
75
+ {%- endif %}
76
+ {{- '\n<tool_response>\n' }}
77
+ {{- content }}
78
+ {{- '\n</tool_response>' }}
79
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
80
+ {{- '<|im_end|>\n' }}
81
+ {%- endif %}
82
+ {%- endif %}
83
+ {%- endfor %}
84
+ {%- if add_generation_prompt %}
85
+ {{- '<|im_start|>assistant\n' }}
86
+ {%- endif %}
checkpoint-105/README.md ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B-Instruct-2507
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - axolotl
7
+ - base_model:adapter:Qwen/Qwen3-4B-Instruct-2507
8
+ - lora
9
+ - transformers
10
+ ---
11
+
12
+ # Model Card for Model ID
13
+
14
+ <!-- Provide a quick summary of what the model is/does. -->
15
+
16
+
17
+
18
+ ## Model Details
19
+
20
+ ### Model Description
21
+
22
+ <!-- Provide a longer summary of what this model is. -->
23
+
24
+
25
+
26
+ - **Developed by:** [More Information Needed]
27
+ - **Funded by [optional]:** [More Information Needed]
28
+ - **Shared by [optional]:** [More Information Needed]
29
+ - **Model type:** [More Information Needed]
30
+ - **Language(s) (NLP):** [More Information Needed]
31
+ - **License:** [More Information Needed]
32
+ - **Finetuned from model [optional]:** [More Information Needed]
33
+
34
+ ### Model Sources [optional]
35
+
36
+ <!-- Provide the basic links for the model. -->
37
+
38
+ - **Repository:** [More Information Needed]
39
+ - **Paper [optional]:** [More Information Needed]
40
+ - **Demo [optional]:** [More Information Needed]
41
+
42
+ ## Uses
43
+
44
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
45
+
46
+ ### Direct Use
47
+
48
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Downstream Use [optional]
53
+
54
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
55
+
56
+ [More Information Needed]
57
+
58
+ ### Out-of-Scope Use
59
+
60
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ## Bias, Risks, and Limitations
65
+
66
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
67
+
68
+ [More Information Needed]
69
+
70
+ ### Recommendations
71
+
72
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
73
+
74
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
75
+
76
+ ## How to Get Started with the Model
77
+
78
+ Use the code below to get started with the model.
79
+
80
+ [More Information Needed]
81
+
82
+ ## Training Details
83
+
84
+ ### Training Data
85
+
86
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
87
+
88
+ [More Information Needed]
89
+
90
+ ### Training Procedure
91
+
92
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
93
+
94
+ #### Preprocessing [optional]
95
+
96
+ [More Information Needed]
97
+
98
+
99
+ #### Training Hyperparameters
100
+
101
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
102
+
103
+ #### Speeds, Sizes, Times [optional]
104
+
105
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
106
+
107
+ [More Information Needed]
108
+
109
+ ## Evaluation
110
+
111
+ <!-- This section describes the evaluation protocols and provides the results. -->
112
+
113
+ ### Testing Data, Factors & Metrics
114
+
115
+ #### Testing Data
116
+
117
+ <!-- This should link to a Dataset Card if possible. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Factors
122
+
123
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
124
+
125
+ [More Information Needed]
126
+
127
+ #### Metrics
128
+
129
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
130
+
131
+ [More Information Needed]
132
+
133
+ ### Results
134
+
135
+ [More Information Needed]
136
+
137
+ #### Summary
138
+
139
+
140
+
141
+ ## Model Examination [optional]
142
+
143
+ <!-- Relevant interpretability work for the model goes here -->
144
+
145
+ [More Information Needed]
146
+
147
+ ## Environmental Impact
148
+
149
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
150
+
151
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
152
+
153
+ - **Hardware Type:** [More Information Needed]
154
+ - **Hours used:** [More Information Needed]
155
+ - **Cloud Provider:** [More Information Needed]
156
+ - **Compute Region:** [More Information Needed]
157
+ - **Carbon Emitted:** [More Information Needed]
158
+
159
+ ## Technical Specifications [optional]
160
+
161
+ ### Model Architecture and Objective
162
+
163
+ [More Information Needed]
164
+
165
+ ### Compute Infrastructure
166
+
167
+ [More Information Needed]
168
+
169
+ #### Hardware
170
+
171
+ [More Information Needed]
172
+
173
+ #### Software
174
+
175
+ [More Information Needed]
176
+
177
+ ## Citation [optional]
178
+
179
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
180
+
181
+ **BibTeX:**
182
+
183
+ [More Information Needed]
184
+
185
+ **APA:**
186
+
187
+ [More Information Needed]
188
+
189
+ ## Glossary [optional]
190
+
191
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
192
+
193
+ [More Information Needed]
194
+
195
+ ## More Information [optional]
196
+
197
+ [More Information Needed]
198
+
199
+ ## Model Card Authors [optional]
200
+
201
+ [More Information Needed]
202
+
203
+ ## Model Card Contact
204
+
205
+ [More Information Needed]
206
+ ### Framework versions
207
+
208
+ - PEFT 0.17.0
checkpoint-105/adapter_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen3-4B-Instruct-2507",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": null,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 8,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "q_proj",
29
+ "gate_proj",
30
+ "up_proj",
31
+ "v_proj",
32
+ "down_proj",
33
+ "k_proj",
34
+ "o_proj"
35
+ ],
36
+ "target_parameters": [],
37
+ "task_type": "CAUSAL_LM",
38
+ "trainable_token_indices": null,
39
+ "use_dora": false,
40
+ "use_qalora": false,
41
+ "use_rslora": false
42
+ }
checkpoint-105/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54d3caf7bf83634e0bf622c9761f1cdc965624c297a13b8d0f45fa42e6a4b3a1
3
+ size 66126768
checkpoint-105/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
checkpoint-105/chat_template.jinja ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if message.content is string %}
27
+ {%- set content = message.content %}
28
+ {%- else %}
29
+ {%- set content = '' %}
30
+ {%- endif %}
31
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
+ {%- elif message.role == "assistant" %}
34
+ {%- set reasoning_content = '' %}
35
+ {%- if message.reasoning_content is string %}
36
+ {%- set reasoning_content = message.reasoning_content %}
37
+ {%- else %}
38
+ {%- if '</think>' in content %}
39
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
+ {%- endif %}
42
+ {%- endif %}
43
+ {%- if loop.index0 > ns.last_query_index %}
44
+ {%- if loop.last or (not loop.last and reasoning_content) %}
45
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
+ {%- else %}
47
+ {{- '<|im_start|>' + message.role + '\n' + content }}
48
+ {%- endif %}
49
+ {%- else %}
50
+ {{- '<|im_start|>' + message.role + '\n' + content }}
51
+ {%- endif %}
52
+ {%- if message.tool_calls %}
53
+ {%- for tool_call in message.tool_calls %}
54
+ {%- if (loop.first and content) or (not loop.first) %}
55
+ {{- '\n' }}
56
+ {%- endif %}
57
+ {%- if tool_call.function %}
58
+ {%- set tool_call = tool_call.function %}
59
+ {%- endif %}
60
+ {{- '<tool_call>\n{"name": "' }}
61
+ {{- tool_call.name }}
62
+ {{- '", "arguments": ' }}
63
+ {%- if tool_call.arguments is string %}
64
+ {{- tool_call.arguments }}
65
+ {%- else %}
66
+ {{- tool_call.arguments | tojson }}
67
+ {%- endif %}
68
+ {{- '}\n</tool_call>' }}
69
+ {%- endfor %}
70
+ {%- endif %}
71
+ {{- '<|im_end|>\n' }}
72
+ {%- elif message.role == "tool" %}
73
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
74
+ {{- '<|im_start|>user' }}
75
+ {%- endif %}
76
+ {{- '\n<tool_response>\n' }}
77
+ {{- content }}
78
+ {{- '\n</tool_response>' }}
79
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
80
+ {{- '<|im_end|>\n' }}
81
+ {%- endif %}
82
+ {%- endif %}
83
+ {%- endfor %}
84
+ {%- if add_generation_prompt %}
85
+ {{- '<|im_start|>assistant\n' }}
86
+ {%- endif %}
checkpoint-105/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-105/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ed27fee10f7b70fc015612c250d7a1cbad38b5bc091082ba56b6843e9687f45
3
+ size 34135860
checkpoint-105/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8573018b24651647d06d6a5ca1df2de350b37825737c4c52dd99d5ae33013b32
3
+ size 14512
checkpoint-105/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4940f16376ab751aa7d868bd930443d7668be27833ea35bde376bcd2d4202ec1
3
+ size 14512
checkpoint-105/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dfcd9279d0def3579f388da83224a4a93c3472451c0ccedd9f00c9689fd66a8
3
+ size 1064
checkpoint-105/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
checkpoint-105/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
checkpoint-105/tokenizer_config.json ADDED
@@ -0,0 +1,239 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|im_end|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 262144,
235
+ "pad_token": "<|endoftext|>",
236
+ "split_special_tokens": false,
237
+ "tokenizer_class": "Qwen2Tokenizer",
238
+ "unk_token": null
239
+ }
checkpoint-105/trainer_state.json ADDED
@@ -0,0 +1,1084 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 105,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.009523809523809525,
14
+ "grad_norm": 2.447430372238159,
15
+ "learning_rate": 0.0,
16
+ "loss": 1.8171,
17
+ "memory/device_mem_reserved(gib)": 36.42,
18
+ "memory/max_mem_active(gib)": 35.17,
19
+ "memory/max_mem_allocated(gib)": 35.17,
20
+ "step": 1
21
+ },
22
+ {
23
+ "epoch": 0.01904761904761905,
24
+ "grad_norm": 2.280487537384033,
25
+ "learning_rate": 2.2222222222222223e-05,
26
+ "loss": 1.7544,
27
+ "memory/device_mem_reserved(gib)": 36.45,
28
+ "memory/max_mem_active(gib)": 35.23,
29
+ "memory/max_mem_allocated(gib)": 35.23,
30
+ "step": 2
31
+ },
32
+ {
33
+ "epoch": 0.02857142857142857,
34
+ "grad_norm": 2.248966693878174,
35
+ "learning_rate": 4.4444444444444447e-05,
36
+ "loss": 1.7546,
37
+ "memory/device_mem_reserved(gib)": 47.1,
38
+ "memory/max_mem_active(gib)": 45.45,
39
+ "memory/max_mem_allocated(gib)": 45.45,
40
+ "step": 3
41
+ },
42
+ {
43
+ "epoch": 0.0380952380952381,
44
+ "grad_norm": 2.2559540271759033,
45
+ "learning_rate": 6.666666666666667e-05,
46
+ "loss": 1.8011,
47
+ "memory/device_mem_reserved(gib)": 47.1,
48
+ "memory/max_mem_active(gib)": 45.45,
49
+ "memory/max_mem_allocated(gib)": 45.45,
50
+ "step": 4
51
+ },
52
+ {
53
+ "epoch": 0.047619047619047616,
54
+ "grad_norm": 2.149872064590454,
55
+ "learning_rate": 8.888888888888889e-05,
56
+ "loss": 1.6465,
57
+ "memory/device_mem_reserved(gib)": 47.1,
58
+ "memory/max_mem_active(gib)": 45.45,
59
+ "memory/max_mem_allocated(gib)": 45.45,
60
+ "step": 5
61
+ },
62
+ {
63
+ "epoch": 0.05714285714285714,
64
+ "grad_norm": 1.6111228466033936,
65
+ "learning_rate": 0.00011111111111111112,
66
+ "loss": 1.5935,
67
+ "memory/device_mem_reserved(gib)": 47.1,
68
+ "memory/max_mem_active(gib)": 45.45,
69
+ "memory/max_mem_allocated(gib)": 45.45,
70
+ "step": 6
71
+ },
72
+ {
73
+ "epoch": 0.06666666666666667,
74
+ "grad_norm": 1.363952875137329,
75
+ "learning_rate": 0.00013333333333333334,
76
+ "loss": 1.4371,
77
+ "memory/device_mem_reserved(gib)": 47.1,
78
+ "memory/max_mem_active(gib)": 45.45,
79
+ "memory/max_mem_allocated(gib)": 45.45,
80
+ "step": 7
81
+ },
82
+ {
83
+ "epoch": 0.0761904761904762,
84
+ "grad_norm": 1.1439425945281982,
85
+ "learning_rate": 0.00015555555555555556,
86
+ "loss": 1.2409,
87
+ "memory/device_mem_reserved(gib)": 47.1,
88
+ "memory/max_mem_active(gib)": 45.45,
89
+ "memory/max_mem_allocated(gib)": 45.45,
90
+ "step": 8
91
+ },
92
+ {
93
+ "epoch": 0.08571428571428572,
94
+ "grad_norm": 0.877252995967865,
95
+ "learning_rate": 0.00017777777777777779,
96
+ "loss": 1.1083,
97
+ "memory/device_mem_reserved(gib)": 47.1,
98
+ "memory/max_mem_active(gib)": 45.45,
99
+ "memory/max_mem_allocated(gib)": 45.45,
100
+ "step": 9
101
+ },
102
+ {
103
+ "epoch": 0.09523809523809523,
104
+ "grad_norm": 0.8071479797363281,
105
+ "learning_rate": 0.0002,
106
+ "loss": 0.9937,
107
+ "memory/device_mem_reserved(gib)": 47.1,
108
+ "memory/max_mem_active(gib)": 45.45,
109
+ "memory/max_mem_allocated(gib)": 45.45,
110
+ "step": 10
111
+ },
112
+ {
113
+ "epoch": 0.10476190476190476,
114
+ "grad_norm": 0.7772606015205383,
115
+ "learning_rate": 0.00019999472984871732,
116
+ "loss": 0.9235,
117
+ "memory/device_mem_reserved(gib)": 47.1,
118
+ "memory/max_mem_active(gib)": 45.45,
119
+ "memory/max_mem_allocated(gib)": 45.45,
120
+ "step": 11
121
+ },
122
+ {
123
+ "epoch": 0.11428571428571428,
124
+ "grad_norm": 0.811764657497406,
125
+ "learning_rate": 0.00019997891995035912,
126
+ "loss": 0.7635,
127
+ "memory/device_mem_reserved(gib)": 47.1,
128
+ "memory/max_mem_active(gib)": 45.45,
129
+ "memory/max_mem_allocated(gib)": 45.45,
130
+ "step": 12
131
+ },
132
+ {
133
+ "epoch": 0.12380952380952381,
134
+ "grad_norm": 0.8423659801483154,
135
+ "learning_rate": 0.0001999525719713366,
136
+ "loss": 0.6836,
137
+ "memory/device_mem_reserved(gib)": 47.1,
138
+ "memory/max_mem_active(gib)": 45.45,
139
+ "memory/max_mem_allocated(gib)": 45.45,
140
+ "step": 13
141
+ },
142
+ {
143
+ "epoch": 0.13333333333333333,
144
+ "grad_norm": 0.7858280539512634,
145
+ "learning_rate": 0.0001999156886888064,
146
+ "loss": 0.6159,
147
+ "memory/device_mem_reserved(gib)": 47.1,
148
+ "memory/max_mem_active(gib)": 45.45,
149
+ "memory/max_mem_allocated(gib)": 45.45,
150
+ "step": 14
151
+ },
152
+ {
153
+ "epoch": 0.14285714285714285,
154
+ "grad_norm": 0.739560604095459,
155
+ "learning_rate": 0.00019986827399037812,
156
+ "loss": 0.5082,
157
+ "memory/device_mem_reserved(gib)": 47.1,
158
+ "memory/max_mem_active(gib)": 45.45,
159
+ "memory/max_mem_allocated(gib)": 45.45,
160
+ "step": 15
161
+ },
162
+ {
163
+ "epoch": 0.1523809523809524,
164
+ "grad_norm": 0.6787500381469727,
165
+ "learning_rate": 0.00019981033287370443,
166
+ "loss": 0.4553,
167
+ "memory/device_mem_reserved(gib)": 47.1,
168
+ "memory/max_mem_active(gib)": 45.45,
169
+ "memory/max_mem_allocated(gib)": 45.45,
170
+ "step": 16
171
+ },
172
+ {
173
+ "epoch": 0.1619047619047619,
174
+ "grad_norm": 0.6258607506752014,
175
+ "learning_rate": 0.00019974187144595432,
176
+ "loss": 0.3913,
177
+ "memory/device_mem_reserved(gib)": 47.1,
178
+ "memory/max_mem_active(gib)": 45.45,
179
+ "memory/max_mem_allocated(gib)": 45.45,
180
+ "step": 17
181
+ },
182
+ {
183
+ "epoch": 0.17142857142857143,
184
+ "grad_norm": 0.47939175367355347,
185
+ "learning_rate": 0.00019966289692316944,
186
+ "loss": 0.3048,
187
+ "memory/device_mem_reserved(gib)": 47.1,
188
+ "memory/max_mem_active(gib)": 45.45,
189
+ "memory/max_mem_allocated(gib)": 45.45,
190
+ "step": 18
191
+ },
192
+ {
193
+ "epoch": 0.18095238095238095,
194
+ "grad_norm": 0.49321815371513367,
195
+ "learning_rate": 0.00019957341762950344,
196
+ "loss": 0.3417,
197
+ "memory/device_mem_reserved(gib)": 47.1,
198
+ "memory/max_mem_active(gib)": 45.45,
199
+ "memory/max_mem_allocated(gib)": 45.45,
200
+ "step": 19
201
+ },
202
+ {
203
+ "epoch": 0.19047619047619047,
204
+ "grad_norm": 0.4252234101295471,
205
+ "learning_rate": 0.00019947344299634464,
206
+ "loss": 0.3033,
207
+ "memory/device_mem_reserved(gib)": 47.1,
208
+ "memory/max_mem_active(gib)": 45.45,
209
+ "memory/max_mem_allocated(gib)": 45.45,
210
+ "step": 20
211
+ },
212
+ {
213
+ "epoch": 0.2,
214
+ "grad_norm": 0.46225014328956604,
215
+ "learning_rate": 0.00019936298356132176,
216
+ "loss": 0.2984,
217
+ "memory/device_mem_reserved(gib)": 47.1,
218
+ "memory/max_mem_active(gib)": 45.45,
219
+ "memory/max_mem_allocated(gib)": 45.45,
220
+ "step": 21
221
+ },
222
+ {
223
+ "epoch": 0.20952380952380953,
224
+ "grad_norm": 0.4106466472148895,
225
+ "learning_rate": 0.0001992420509671936,
226
+ "loss": 0.2764,
227
+ "memory/device_mem_reserved(gib)": 47.12,
228
+ "memory/max_mem_active(gib)": 45.46,
229
+ "memory/max_mem_allocated(gib)": 45.46,
230
+ "step": 22
231
+ },
232
+ {
233
+ "epoch": 0.21904761904761905,
234
+ "grad_norm": 0.3764067590236664,
235
+ "learning_rate": 0.00019911065796062135,
236
+ "loss": 0.2352,
237
+ "memory/device_mem_reserved(gib)": 47.12,
238
+ "memory/max_mem_active(gib)": 45.46,
239
+ "memory/max_mem_allocated(gib)": 45.46,
240
+ "step": 23
241
+ },
242
+ {
243
+ "epoch": 0.22857142857142856,
244
+ "grad_norm": 0.38518962264060974,
245
+ "learning_rate": 0.00019896881839082556,
246
+ "loss": 0.2239,
247
+ "memory/device_mem_reserved(gib)": 47.12,
248
+ "memory/max_mem_active(gib)": 45.46,
249
+ "memory/max_mem_allocated(gib)": 45.46,
250
+ "step": 24
251
+ },
252
+ {
253
+ "epoch": 0.23809523809523808,
254
+ "grad_norm": 0.39728403091430664,
255
+ "learning_rate": 0.00019881654720812594,
256
+ "loss": 0.2192,
257
+ "memory/device_mem_reserved(gib)": 47.13,
258
+ "memory/max_mem_active(gib)": 45.46,
259
+ "memory/max_mem_allocated(gib)": 45.46,
260
+ "step": 25
261
+ },
262
+ {
263
+ "epoch": 0.24761904761904763,
264
+ "grad_norm": 0.3667093813419342,
265
+ "learning_rate": 0.00019865386046236596,
266
+ "loss": 0.2834,
267
+ "memory/device_mem_reserved(gib)": 47.13,
268
+ "memory/max_mem_active(gib)": 45.46,
269
+ "memory/max_mem_allocated(gib)": 45.46,
270
+ "step": 26
271
+ },
272
+ {
273
+ "epoch": 0.2571428571428571,
274
+ "grad_norm": 0.32502660155296326,
275
+ "learning_rate": 0.00019848077530122083,
276
+ "loss": 0.2166,
277
+ "memory/device_mem_reserved(gib)": 47.13,
278
+ "memory/max_mem_active(gib)": 45.46,
279
+ "memory/max_mem_allocated(gib)": 45.46,
280
+ "step": 27
281
+ },
282
+ {
283
+ "epoch": 0.26666666666666666,
284
+ "grad_norm": 0.31225672364234924,
285
+ "learning_rate": 0.0001982973099683902,
286
+ "loss": 0.2238,
287
+ "memory/device_mem_reserved(gib)": 47.13,
288
+ "memory/max_mem_active(gib)": 45.46,
289
+ "memory/max_mem_allocated(gib)": 45.46,
290
+ "step": 28
291
+ },
292
+ {
293
+ "epoch": 0.2761904761904762,
294
+ "grad_norm": 0.27448731660842896,
295
+ "learning_rate": 0.00019810348380167527,
296
+ "loss": 0.1983,
297
+ "memory/device_mem_reserved(gib)": 47.13,
298
+ "memory/max_mem_active(gib)": 45.46,
299
+ "memory/max_mem_allocated(gib)": 45.46,
300
+ "step": 29
301
+ },
302
+ {
303
+ "epoch": 0.2857142857142857,
304
+ "grad_norm": 0.32014158368110657,
305
+ "learning_rate": 0.00019789931723094046,
306
+ "loss": 0.2554,
307
+ "memory/device_mem_reserved(gib)": 47.13,
308
+ "memory/max_mem_active(gib)": 45.46,
309
+ "memory/max_mem_allocated(gib)": 45.46,
310
+ "step": 30
311
+ },
312
+ {
313
+ "epoch": 0.29523809523809524,
314
+ "grad_norm": 0.39480143785476685,
315
+ "learning_rate": 0.0001976848317759601,
316
+ "loss": 0.2438,
317
+ "memory/device_mem_reserved(gib)": 47.13,
318
+ "memory/max_mem_active(gib)": 45.46,
319
+ "memory/max_mem_allocated(gib)": 45.46,
320
+ "step": 31
321
+ },
322
+ {
323
+ "epoch": 0.3047619047619048,
324
+ "grad_norm": 0.28004321455955505,
325
+ "learning_rate": 0.00019746005004415005,
326
+ "loss": 0.2145,
327
+ "memory/device_mem_reserved(gib)": 47.13,
328
+ "memory/max_mem_active(gib)": 45.46,
329
+ "memory/max_mem_allocated(gib)": 45.46,
330
+ "step": 32
331
+ },
332
+ {
333
+ "epoch": 0.3142857142857143,
334
+ "grad_norm": 0.2754361033439636,
335
+ "learning_rate": 0.00019722499572818496,
336
+ "loss": 0.2284,
337
+ "memory/device_mem_reserved(gib)": 47.13,
338
+ "memory/max_mem_active(gib)": 45.46,
339
+ "memory/max_mem_allocated(gib)": 45.46,
340
+ "step": 33
341
+ },
342
+ {
343
+ "epoch": 0.3238095238095238,
344
+ "grad_norm": 0.3160915970802307,
345
+ "learning_rate": 0.00019697969360350098,
346
+ "loss": 0.2482,
347
+ "memory/device_mem_reserved(gib)": 47.13,
348
+ "memory/max_mem_active(gib)": 45.47,
349
+ "memory/max_mem_allocated(gib)": 45.47,
350
+ "step": 34
351
+ },
352
+ {
353
+ "epoch": 0.3333333333333333,
354
+ "grad_norm": 0.24699093401432037,
355
+ "learning_rate": 0.00019672416952568416,
356
+ "loss": 0.1916,
357
+ "memory/device_mem_reserved(gib)": 47.13,
358
+ "memory/max_mem_active(gib)": 45.47,
359
+ "memory/max_mem_allocated(gib)": 45.47,
360
+ "step": 35
361
+ },
362
+ {
363
+ "epoch": 0.34285714285714286,
364
+ "grad_norm": 0.2674485743045807,
365
+ "learning_rate": 0.00019645845042774553,
366
+ "loss": 0.2204,
367
+ "memory/device_mem_reserved(gib)": 47.13,
368
+ "memory/max_mem_active(gib)": 45.47,
369
+ "memory/max_mem_allocated(gib)": 45.47,
370
+ "step": 36
371
+ },
372
+ {
373
+ "epoch": 0.3523809523809524,
374
+ "grad_norm": 0.33608749508857727,
375
+ "learning_rate": 0.00019618256431728194,
376
+ "loss": 0.235,
377
+ "memory/device_mem_reserved(gib)": 47.13,
378
+ "memory/max_mem_active(gib)": 45.47,
379
+ "memory/max_mem_allocated(gib)": 45.47,
380
+ "step": 37
381
+ },
382
+ {
383
+ "epoch": 0.3619047619047619,
384
+ "grad_norm": 0.29581594467163086,
385
+ "learning_rate": 0.00019589654027352414,
386
+ "loss": 0.2175,
387
+ "memory/device_mem_reserved(gib)": 47.13,
388
+ "memory/max_mem_active(gib)": 45.47,
389
+ "memory/max_mem_allocated(gib)": 45.47,
390
+ "step": 38
391
+ },
392
+ {
393
+ "epoch": 0.37142857142857144,
394
+ "grad_norm": 0.3526640832424164,
395
+ "learning_rate": 0.0001956004084442718,
396
+ "loss": 0.1989,
397
+ "memory/device_mem_reserved(gib)": 47.13,
398
+ "memory/max_mem_active(gib)": 45.47,
399
+ "memory/max_mem_allocated(gib)": 45.47,
400
+ "step": 39
401
+ },
402
+ {
403
+ "epoch": 0.38095238095238093,
404
+ "grad_norm": 0.2882135808467865,
405
+ "learning_rate": 0.00019529420004271567,
406
+ "loss": 0.2144,
407
+ "memory/device_mem_reserved(gib)": 47.13,
408
+ "memory/max_mem_active(gib)": 45.47,
409
+ "memory/max_mem_allocated(gib)": 45.47,
410
+ "step": 40
411
+ },
412
+ {
413
+ "epoch": 0.3904761904761905,
414
+ "grad_norm": 0.2675183117389679,
415
+ "learning_rate": 0.0001949779473441478,
416
+ "loss": 0.1987,
417
+ "memory/device_mem_reserved(gib)": 47.13,
418
+ "memory/max_mem_active(gib)": 45.47,
419
+ "memory/max_mem_allocated(gib)": 45.47,
420
+ "step": 41
421
+ },
422
+ {
423
+ "epoch": 0.4,
424
+ "grad_norm": 0.2908264696598053,
425
+ "learning_rate": 0.00019465168368255946,
426
+ "loss": 0.1893,
427
+ "memory/device_mem_reserved(gib)": 47.13,
428
+ "memory/max_mem_active(gib)": 45.47,
429
+ "memory/max_mem_allocated(gib)": 45.47,
430
+ "step": 42
431
+ },
432
+ {
433
+ "epoch": 0.4095238095238095,
434
+ "grad_norm": 0.3211234211921692,
435
+ "learning_rate": 0.00019431544344712776,
436
+ "loss": 0.2008,
437
+ "memory/device_mem_reserved(gib)": 47.13,
438
+ "memory/max_mem_active(gib)": 45.47,
439
+ "memory/max_mem_allocated(gib)": 45.47,
440
+ "step": 43
441
+ },
442
+ {
443
+ "epoch": 0.41904761904761906,
444
+ "grad_norm": 0.2795293927192688,
445
+ "learning_rate": 0.00019396926207859084,
446
+ "loss": 0.1935,
447
+ "memory/device_mem_reserved(gib)": 47.13,
448
+ "memory/max_mem_active(gib)": 45.47,
449
+ "memory/max_mem_allocated(gib)": 45.47,
450
+ "step": 44
451
+ },
452
+ {
453
+ "epoch": 0.42857142857142855,
454
+ "grad_norm": 0.28679871559143066,
455
+ "learning_rate": 0.00019361317606551238,
456
+ "loss": 0.1987,
457
+ "memory/device_mem_reserved(gib)": 47.13,
458
+ "memory/max_mem_active(gib)": 45.47,
459
+ "memory/max_mem_allocated(gib)": 45.47,
460
+ "step": 45
461
+ },
462
+ {
463
+ "epoch": 0.4380952380952381,
464
+ "grad_norm": 0.3165067732334137,
465
+ "learning_rate": 0.00019324722294043558,
466
+ "loss": 0.2118,
467
+ "memory/device_mem_reserved(gib)": 47.13,
468
+ "memory/max_mem_active(gib)": 45.47,
469
+ "memory/max_mem_allocated(gib)": 45.47,
470
+ "step": 46
471
+ },
472
+ {
473
+ "epoch": 0.44761904761904764,
474
+ "grad_norm": 0.25720369815826416,
475
+ "learning_rate": 0.00019287144127592704,
476
+ "loss": 0.1962,
477
+ "memory/device_mem_reserved(gib)": 47.13,
478
+ "memory/max_mem_active(gib)": 45.47,
479
+ "memory/max_mem_allocated(gib)": 45.47,
480
+ "step": 47
481
+ },
482
+ {
483
+ "epoch": 0.45714285714285713,
484
+ "grad_norm": 0.2577202022075653,
485
+ "learning_rate": 0.0001924858706805112,
486
+ "loss": 0.2022,
487
+ "memory/device_mem_reserved(gib)": 47.13,
488
+ "memory/max_mem_active(gib)": 45.47,
489
+ "memory/max_mem_allocated(gib)": 45.47,
490
+ "step": 48
491
+ },
492
+ {
493
+ "epoch": 0.4666666666666667,
494
+ "grad_norm": 0.2605520188808441,
495
+ "learning_rate": 0.0001920905517944954,
496
+ "loss": 0.1891,
497
+ "memory/device_mem_reserved(gib)": 47.13,
498
+ "memory/max_mem_active(gib)": 45.47,
499
+ "memory/max_mem_allocated(gib)": 45.47,
500
+ "step": 49
501
+ },
502
+ {
503
+ "epoch": 0.47619047619047616,
504
+ "grad_norm": 0.2890350818634033,
505
+ "learning_rate": 0.00019168552628568631,
506
+ "loss": 0.17,
507
+ "memory/device_mem_reserved(gib)": 47.13,
508
+ "memory/max_mem_active(gib)": 45.47,
509
+ "memory/max_mem_allocated(gib)": 45.47,
510
+ "step": 50
511
+ },
512
+ {
513
+ "epoch": 0.4857142857142857,
514
+ "grad_norm": 0.25532475113868713,
515
+ "learning_rate": 0.00019127083684499806,
516
+ "loss": 0.1796,
517
+ "memory/device_mem_reserved(gib)": 47.13,
518
+ "memory/max_mem_active(gib)": 45.47,
519
+ "memory/max_mem_allocated(gib)": 45.47,
520
+ "step": 51
521
+ },
522
+ {
523
+ "epoch": 0.49523809523809526,
524
+ "grad_norm": 0.2963904142379761,
525
+ "learning_rate": 0.00019084652718195238,
526
+ "loss": 0.19,
527
+ "memory/device_mem_reserved(gib)": 47.13,
528
+ "memory/max_mem_active(gib)": 45.47,
529
+ "memory/max_mem_allocated(gib)": 45.47,
530
+ "step": 52
531
+ },
532
+ {
533
+ "epoch": 0.5047619047619047,
534
+ "grad_norm": 0.24455095827579498,
535
+ "learning_rate": 0.0001904126420200716,
536
+ "loss": 0.2017,
537
+ "memory/device_mem_reserved(gib)": 47.13,
538
+ "memory/max_mem_active(gib)": 45.47,
539
+ "memory/max_mem_allocated(gib)": 45.47,
540
+ "step": 53
541
+ },
542
+ {
543
+ "epoch": 0.5142857142857142,
544
+ "grad_norm": 0.2518145442008972,
545
+ "learning_rate": 0.00018996922709216455,
546
+ "loss": 0.1896,
547
+ "memory/device_mem_reserved(gib)": 47.13,
548
+ "memory/max_mem_active(gib)": 45.47,
549
+ "memory/max_mem_allocated(gib)": 45.47,
550
+ "step": 54
551
+ },
552
+ {
553
+ "epoch": 0.5238095238095238,
554
+ "grad_norm": 0.26225805282592773,
555
+ "learning_rate": 0.00018951632913550626,
556
+ "loss": 0.1893,
557
+ "memory/device_mem_reserved(gib)": 47.13,
558
+ "memory/max_mem_active(gib)": 45.47,
559
+ "memory/max_mem_allocated(gib)": 45.47,
560
+ "step": 55
561
+ },
562
+ {
563
+ "epoch": 0.5333333333333333,
564
+ "grad_norm": 0.25169169902801514,
565
+ "learning_rate": 0.00018905399588691163,
566
+ "loss": 0.1939,
567
+ "memory/device_mem_reserved(gib)": 47.13,
568
+ "memory/max_mem_active(gib)": 45.47,
569
+ "memory/max_mem_allocated(gib)": 45.47,
570
+ "step": 56
571
+ },
572
+ {
573
+ "epoch": 0.5428571428571428,
574
+ "grad_norm": 0.23847182095050812,
575
+ "learning_rate": 0.00018858227607770398,
576
+ "loss": 0.1686,
577
+ "memory/device_mem_reserved(gib)": 47.13,
578
+ "memory/max_mem_active(gib)": 45.47,
579
+ "memory/max_mem_allocated(gib)": 45.47,
580
+ "step": 57
581
+ },
582
+ {
583
+ "epoch": 0.5523809523809524,
584
+ "grad_norm": 0.30888280272483826,
585
+ "learning_rate": 0.00018810121942857845,
586
+ "loss": 0.1768,
587
+ "memory/device_mem_reserved(gib)": 47.13,
588
+ "memory/max_mem_active(gib)": 45.47,
589
+ "memory/max_mem_allocated(gib)": 45.47,
590
+ "step": 58
591
+ },
592
+ {
593
+ "epoch": 0.5619047619047619,
594
+ "grad_norm": 0.27275580167770386,
595
+ "learning_rate": 0.00018761087664436138,
596
+ "loss": 0.1941,
597
+ "memory/device_mem_reserved(gib)": 47.13,
598
+ "memory/max_mem_active(gib)": 45.47,
599
+ "memory/max_mem_allocated(gib)": 45.47,
600
+ "step": 59
601
+ },
602
+ {
603
+ "epoch": 0.5714285714285714,
604
+ "grad_norm": 0.22851359844207764,
605
+ "learning_rate": 0.00018711129940866575,
606
+ "loss": 0.1658,
607
+ "memory/device_mem_reserved(gib)": 47.13,
608
+ "memory/max_mem_active(gib)": 45.47,
609
+ "memory/max_mem_allocated(gib)": 45.47,
610
+ "step": 60
611
+ },
612
+ {
613
+ "epoch": 0.580952380952381,
614
+ "grad_norm": 0.25515425205230713,
615
+ "learning_rate": 0.00018660254037844388,
616
+ "loss": 0.1858,
617
+ "memory/device_mem_reserved(gib)": 47.13,
618
+ "memory/max_mem_active(gib)": 45.47,
619
+ "memory/max_mem_allocated(gib)": 45.47,
620
+ "step": 61
621
+ },
622
+ {
623
+ "epoch": 0.5904761904761905,
624
+ "grad_norm": 0.24301765859127045,
625
+ "learning_rate": 0.00018608465317843678,
626
+ "loss": 0.1892,
627
+ "memory/device_mem_reserved(gib)": 47.13,
628
+ "memory/max_mem_active(gib)": 45.47,
629
+ "memory/max_mem_allocated(gib)": 45.47,
630
+ "step": 62
631
+ },
632
+ {
633
+ "epoch": 0.6,
634
+ "grad_norm": 0.24329794943332672,
635
+ "learning_rate": 0.00018555769239552233,
636
+ "loss": 0.1719,
637
+ "memory/device_mem_reserved(gib)": 47.13,
638
+ "memory/max_mem_active(gib)": 45.47,
639
+ "memory/max_mem_allocated(gib)": 45.47,
640
+ "step": 63
641
+ },
642
+ {
643
+ "epoch": 0.6095238095238096,
644
+ "grad_norm": 0.3206610381603241,
645
+ "learning_rate": 0.00018502171357296144,
646
+ "loss": 0.2075,
647
+ "memory/device_mem_reserved(gib)": 47.13,
648
+ "memory/max_mem_active(gib)": 45.47,
649
+ "memory/max_mem_allocated(gib)": 45.47,
650
+ "step": 64
651
+ },
652
+ {
653
+ "epoch": 0.6190476190476191,
654
+ "grad_norm": 0.3089618980884552,
655
+ "learning_rate": 0.00018447677320454367,
656
+ "loss": 0.185,
657
+ "memory/device_mem_reserved(gib)": 47.13,
658
+ "memory/max_mem_active(gib)": 45.47,
659
+ "memory/max_mem_allocated(gib)": 45.47,
660
+ "step": 65
661
+ },
662
+ {
663
+ "epoch": 0.6285714285714286,
664
+ "grad_norm": 0.291916161775589,
665
+ "learning_rate": 0.00018392292872863267,
666
+ "loss": 0.1739,
667
+ "memory/device_mem_reserved(gib)": 47.13,
668
+ "memory/max_mem_active(gib)": 45.47,
669
+ "memory/max_mem_allocated(gib)": 45.47,
670
+ "step": 66
671
+ },
672
+ {
673
+ "epoch": 0.638095238095238,
674
+ "grad_norm": 0.25783228874206543,
675
+ "learning_rate": 0.00018336023852211195,
676
+ "loss": 0.2061,
677
+ "memory/device_mem_reserved(gib)": 47.13,
678
+ "memory/max_mem_active(gib)": 45.47,
679
+ "memory/max_mem_allocated(gib)": 45.47,
680
+ "step": 67
681
+ },
682
+ {
683
+ "epoch": 0.6476190476190476,
684
+ "grad_norm": 0.310614675283432,
685
+ "learning_rate": 0.00018278876189423179,
686
+ "loss": 0.1984,
687
+ "memory/device_mem_reserved(gib)": 47.13,
688
+ "memory/max_mem_active(gib)": 45.47,
689
+ "memory/max_mem_allocated(gib)": 45.47,
690
+ "step": 68
691
+ },
692
+ {
693
+ "epoch": 0.6571428571428571,
694
+ "grad_norm": 0.23836325109004974,
695
+ "learning_rate": 0.00018220855908035783,
696
+ "loss": 0.1742,
697
+ "memory/device_mem_reserved(gib)": 47.13,
698
+ "memory/max_mem_active(gib)": 45.47,
699
+ "memory/max_mem_allocated(gib)": 45.47,
700
+ "step": 69
701
+ },
702
+ {
703
+ "epoch": 0.6666666666666666,
704
+ "grad_norm": 0.30577945709228516,
705
+ "learning_rate": 0.0001816196912356222,
706
+ "loss": 0.1954,
707
+ "memory/device_mem_reserved(gib)": 47.13,
708
+ "memory/max_mem_active(gib)": 45.47,
709
+ "memory/max_mem_allocated(gib)": 45.47,
710
+ "step": 70
711
+ },
712
+ {
713
+ "epoch": 0.6761904761904762,
714
+ "grad_norm": 0.3177741467952728,
715
+ "learning_rate": 0.00018102222042847737,
716
+ "loss": 0.193,
717
+ "memory/device_mem_reserved(gib)": 47.13,
718
+ "memory/max_mem_active(gib)": 45.47,
719
+ "memory/max_mem_allocated(gib)": 45.47,
720
+ "step": 71
721
+ },
722
+ {
723
+ "epoch": 0.6857142857142857,
724
+ "grad_norm": 0.3444816470146179,
725
+ "learning_rate": 0.00018041620963415417,
726
+ "loss": 0.1837,
727
+ "memory/device_mem_reserved(gib)": 47.13,
728
+ "memory/max_mem_active(gib)": 45.47,
729
+ "memory/max_mem_allocated(gib)": 45.47,
730
+ "step": 72
731
+ },
732
+ {
733
+ "epoch": 0.6952380952380952,
734
+ "grad_norm": 0.3139638900756836,
735
+ "learning_rate": 0.000179801722728024,
736
+ "loss": 0.1902,
737
+ "memory/device_mem_reserved(gib)": 47.13,
738
+ "memory/max_mem_active(gib)": 45.47,
739
+ "memory/max_mem_allocated(gib)": 45.47,
740
+ "step": 73
741
+ },
742
+ {
743
+ "epoch": 0.7047619047619048,
744
+ "grad_norm": 0.24754184484481812,
745
+ "learning_rate": 0.00017917882447886582,
746
+ "loss": 0.1612,
747
+ "memory/device_mem_reserved(gib)": 47.13,
748
+ "memory/max_mem_active(gib)": 45.47,
749
+ "memory/max_mem_allocated(gib)": 45.47,
750
+ "step": 74
751
+ },
752
+ {
753
+ "epoch": 0.7142857142857143,
754
+ "grad_norm": 0.23204036056995392,
755
+ "learning_rate": 0.00017854758054203988,
756
+ "loss": 0.1694,
757
+ "memory/device_mem_reserved(gib)": 47.13,
758
+ "memory/max_mem_active(gib)": 45.47,
759
+ "memory/max_mem_allocated(gib)": 45.47,
760
+ "step": 75
761
+ },
762
+ {
763
+ "epoch": 0.7238095238095238,
764
+ "grad_norm": 0.2823260426521301,
765
+ "learning_rate": 0.00017790805745256704,
766
+ "loss": 0.1739,
767
+ "memory/device_mem_reserved(gib)": 47.13,
768
+ "memory/max_mem_active(gib)": 45.47,
769
+ "memory/max_mem_allocated(gib)": 45.47,
770
+ "step": 76
771
+ },
772
+ {
773
+ "epoch": 0.7333333333333333,
774
+ "grad_norm": 0.3193572163581848,
775
+ "learning_rate": 0.0001772603226181159,
776
+ "loss": 0.1924,
777
+ "memory/device_mem_reserved(gib)": 47.13,
778
+ "memory/max_mem_active(gib)": 45.47,
779
+ "memory/max_mem_allocated(gib)": 45.47,
780
+ "step": 77
781
+ },
782
+ {
783
+ "epoch": 0.7428571428571429,
784
+ "grad_norm": 0.30530717968940735,
785
+ "learning_rate": 0.0001766044443118978,
786
+ "loss": 0.1375,
787
+ "memory/device_mem_reserved(gib)": 47.13,
788
+ "memory/max_mem_active(gib)": 45.47,
789
+ "memory/max_mem_allocated(gib)": 45.47,
790
+ "step": 78
791
+ },
792
+ {
793
+ "epoch": 0.7523809523809524,
794
+ "grad_norm": 0.2974378764629364,
795
+ "learning_rate": 0.00017594049166547073,
796
+ "loss": 0.1866,
797
+ "memory/device_mem_reserved(gib)": 47.13,
798
+ "memory/max_mem_active(gib)": 45.47,
799
+ "memory/max_mem_allocated(gib)": 45.47,
800
+ "step": 79
801
+ },
802
+ {
803
+ "epoch": 0.7619047619047619,
804
+ "grad_norm": 0.26553574204444885,
805
+ "learning_rate": 0.00017526853466145244,
806
+ "loss": 0.193,
807
+ "memory/device_mem_reserved(gib)": 47.13,
808
+ "memory/max_mem_active(gib)": 45.47,
809
+ "memory/max_mem_allocated(gib)": 45.47,
810
+ "step": 80
811
+ },
812
+ {
813
+ "epoch": 0.7714285714285715,
814
+ "grad_norm": 0.32059594988822937,
815
+ "learning_rate": 0.00017458864412614434,
816
+ "loss": 0.1822,
817
+ "memory/device_mem_reserved(gib)": 47.13,
818
+ "memory/max_mem_active(gib)": 45.47,
819
+ "memory/max_mem_allocated(gib)": 45.47,
820
+ "step": 81
821
+ },
822
+ {
823
+ "epoch": 0.780952380952381,
824
+ "grad_norm": 0.3260822594165802,
825
+ "learning_rate": 0.00017390089172206592,
826
+ "loss": 0.1734,
827
+ "memory/device_mem_reserved(gib)": 47.13,
828
+ "memory/max_mem_active(gib)": 45.47,
829
+ "memory/max_mem_allocated(gib)": 45.47,
830
+ "step": 82
831
+ },
832
+ {
833
+ "epoch": 0.7904761904761904,
834
+ "grad_norm": 0.2754499912261963,
835
+ "learning_rate": 0.00017320534994040148,
836
+ "loss": 0.1814,
837
+ "memory/device_mem_reserved(gib)": 47.13,
838
+ "memory/max_mem_active(gib)": 45.47,
839
+ "memory/max_mem_allocated(gib)": 45.47,
840
+ "step": 83
841
+ },
842
+ {
843
+ "epoch": 0.8,
844
+ "grad_norm": 0.2639872431755066,
845
+ "learning_rate": 0.00017250209209335927,
846
+ "loss": 0.1764,
847
+ "memory/device_mem_reserved(gib)": 47.13,
848
+ "memory/max_mem_active(gib)": 45.47,
849
+ "memory/max_mem_allocated(gib)": 45.47,
850
+ "step": 84
851
+ },
852
+ {
853
+ "epoch": 0.8095238095238095,
854
+ "grad_norm": 0.22905386984348297,
855
+ "learning_rate": 0.0001717911923064442,
856
+ "loss": 0.1365,
857
+ "memory/device_mem_reserved(gib)": 47.13,
858
+ "memory/max_mem_active(gib)": 45.47,
859
+ "memory/max_mem_allocated(gib)": 45.47,
860
+ "step": 85
861
+ },
862
+ {
863
+ "epoch": 0.819047619047619,
864
+ "grad_norm": 0.26005449891090393,
865
+ "learning_rate": 0.00017107272551064473,
866
+ "loss": 0.1756,
867
+ "memory/device_mem_reserved(gib)": 47.13,
868
+ "memory/max_mem_active(gib)": 45.47,
869
+ "memory/max_mem_allocated(gib)": 45.47,
870
+ "step": 86
871
+ },
872
+ {
873
+ "epoch": 0.8285714285714286,
874
+ "grad_norm": 0.25042417645454407,
875
+ "learning_rate": 0.00017034676743453499,
876
+ "loss": 0.15,
877
+ "memory/device_mem_reserved(gib)": 47.13,
878
+ "memory/max_mem_active(gib)": 45.47,
879
+ "memory/max_mem_allocated(gib)": 45.47,
880
+ "step": 87
881
+ },
882
+ {
883
+ "epoch": 0.8380952380952381,
884
+ "grad_norm": 0.34358277916908264,
885
+ "learning_rate": 0.0001696133945962927,
886
+ "loss": 0.1763,
887
+ "memory/device_mem_reserved(gib)": 47.13,
888
+ "memory/max_mem_active(gib)": 45.47,
889
+ "memory/max_mem_allocated(gib)": 45.47,
890
+ "step": 88
891
+ },
892
+ {
893
+ "epoch": 0.8476190476190476,
894
+ "grad_norm": 0.23855257034301758,
895
+ "learning_rate": 0.0001688726842956339,
896
+ "loss": 0.1547,
897
+ "memory/device_mem_reserved(gib)": 47.13,
898
+ "memory/max_mem_active(gib)": 45.47,
899
+ "memory/max_mem_allocated(gib)": 45.47,
900
+ "step": 89
901
+ },
902
+ {
903
+ "epoch": 0.8571428571428571,
904
+ "grad_norm": 0.2579350173473358,
905
+ "learning_rate": 0.0001681247146056654,
906
+ "loss": 0.1872,
907
+ "memory/device_mem_reserved(gib)": 47.13,
908
+ "memory/max_mem_active(gib)": 45.47,
909
+ "memory/max_mem_allocated(gib)": 45.47,
910
+ "step": 90
911
+ },
912
+ {
913
+ "epoch": 0.8666666666666667,
914
+ "grad_norm": 0.30034390091896057,
915
+ "learning_rate": 0.00016736956436465573,
916
+ "loss": 0.1558,
917
+ "memory/device_mem_reserved(gib)": 47.13,
918
+ "memory/max_mem_active(gib)": 45.47,
919
+ "memory/max_mem_allocated(gib)": 45.47,
920
+ "step": 91
921
+ },
922
+ {
923
+ "epoch": 0.8761904761904762,
924
+ "grad_norm": 0.28185054659843445,
925
+ "learning_rate": 0.00016660731316772505,
926
+ "loss": 0.1878,
927
+ "memory/device_mem_reserved(gib)": 47.13,
928
+ "memory/max_mem_active(gib)": 45.47,
929
+ "memory/max_mem_allocated(gib)": 45.47,
930
+ "step": 92
931
+ },
932
+ {
933
+ "epoch": 0.8857142857142857,
934
+ "grad_norm": 0.32514333724975586,
935
+ "learning_rate": 0.0001658380413584558,
936
+ "loss": 0.1705,
937
+ "memory/device_mem_reserved(gib)": 47.13,
938
+ "memory/max_mem_active(gib)": 45.47,
939
+ "memory/max_mem_allocated(gib)": 45.47,
940
+ "step": 93
941
+ },
942
+ {
943
+ "epoch": 0.8952380952380953,
944
+ "grad_norm": 0.30903860926628113,
945
+ "learning_rate": 0.0001650618300204242,
946
+ "loss": 0.1708,
947
+ "memory/device_mem_reserved(gib)": 47.13,
948
+ "memory/max_mem_active(gib)": 45.47,
949
+ "memory/max_mem_allocated(gib)": 45.47,
950
+ "step": 94
951
+ },
952
+ {
953
+ "epoch": 0.9047619047619048,
954
+ "grad_norm": 0.24723806977272034,
955
+ "learning_rate": 0.00016427876096865394,
956
+ "loss": 0.1571,
957
+ "memory/device_mem_reserved(gib)": 47.13,
958
+ "memory/max_mem_active(gib)": 45.47,
959
+ "memory/max_mem_allocated(gib)": 45.47,
960
+ "step": 95
961
+ },
962
+ {
963
+ "epoch": 0.9142857142857143,
964
+ "grad_norm": 0.2727813422679901,
965
+ "learning_rate": 0.0001634889167409923,
966
+ "loss": 0.1779,
967
+ "memory/device_mem_reserved(gib)": 47.13,
968
+ "memory/max_mem_active(gib)": 45.47,
969
+ "memory/max_mem_allocated(gib)": 45.47,
970
+ "step": 96
971
+ },
972
+ {
973
+ "epoch": 0.9238095238095239,
974
+ "grad_norm": 0.2521083354949951,
975
+ "learning_rate": 0.0001626923805894107,
976
+ "loss": 0.1451,
977
+ "memory/device_mem_reserved(gib)": 47.13,
978
+ "memory/max_mem_active(gib)": 45.47,
979
+ "memory/max_mem_allocated(gib)": 45.47,
980
+ "step": 97
981
+ },
982
+ {
983
+ "epoch": 0.9333333333333333,
984
+ "grad_norm": 0.25517475605010986,
985
+ "learning_rate": 0.00016188923647122947,
986
+ "loss": 0.1789,
987
+ "memory/device_mem_reserved(gib)": 47.13,
988
+ "memory/max_mem_active(gib)": 45.47,
989
+ "memory/max_mem_allocated(gib)": 45.47,
990
+ "step": 98
991
+ },
992
+ {
993
+ "epoch": 0.9428571428571428,
994
+ "grad_norm": 0.26679396629333496,
995
+ "learning_rate": 0.0001610795690402688,
996
+ "loss": 0.1835,
997
+ "memory/device_mem_reserved(gib)": 47.13,
998
+ "memory/max_mem_active(gib)": 45.47,
999
+ "memory/max_mem_allocated(gib)": 45.47,
1000
+ "step": 99
1001
+ },
1002
+ {
1003
+ "epoch": 0.9523809523809523,
1004
+ "grad_norm": 0.22741852700710297,
1005
+ "learning_rate": 0.00016026346363792567,
1006
+ "loss": 0.1599,
1007
+ "memory/device_mem_reserved(gib)": 47.13,
1008
+ "memory/max_mem_active(gib)": 45.47,
1009
+ "memory/max_mem_allocated(gib)": 45.47,
1010
+ "step": 100
1011
+ },
1012
+ {
1013
+ "epoch": 0.9619047619047619,
1014
+ "grad_norm": 0.28691986203193665,
1015
+ "learning_rate": 0.00015944100628417868,
1016
+ "loss": 0.1654,
1017
+ "memory/device_mem_reserved(gib)": 47.13,
1018
+ "memory/max_mem_active(gib)": 45.47,
1019
+ "memory/max_mem_allocated(gib)": 45.47,
1020
+ "step": 101
1021
+ },
1022
+ {
1023
+ "epoch": 0.9714285714285714,
1024
+ "grad_norm": 0.24840302765369415,
1025
+ "learning_rate": 0.00015861228366852148,
1026
+ "loss": 0.1584,
1027
+ "memory/device_mem_reserved(gib)": 47.13,
1028
+ "memory/max_mem_active(gib)": 45.47,
1029
+ "memory/max_mem_allocated(gib)": 45.47,
1030
+ "step": 102
1031
+ },
1032
+ {
1033
+ "epoch": 0.9809523809523809,
1034
+ "grad_norm": 0.24804599583148956,
1035
+ "learning_rate": 0.00015777738314082514,
1036
+ "loss": 0.1391,
1037
+ "memory/device_mem_reserved(gib)": 47.13,
1038
+ "memory/max_mem_active(gib)": 45.47,
1039
+ "memory/max_mem_allocated(gib)": 45.47,
1040
+ "step": 103
1041
+ },
1042
+ {
1043
+ "epoch": 0.9904761904761905,
1044
+ "grad_norm": 0.2463735193014145,
1045
+ "learning_rate": 0.00015693639270213136,
1046
+ "loss": 0.1408,
1047
+ "memory/device_mem_reserved(gib)": 47.13,
1048
+ "memory/max_mem_active(gib)": 45.47,
1049
+ "memory/max_mem_allocated(gib)": 45.47,
1050
+ "step": 104
1051
+ },
1052
+ {
1053
+ "epoch": 1.0,
1054
+ "grad_norm": 0.29896217584609985,
1055
+ "learning_rate": 0.000156089400995377,
1056
+ "loss": 0.1666,
1057
+ "memory/device_mem_reserved(gib)": 47.13,
1058
+ "memory/max_mem_active(gib)": 45.47,
1059
+ "memory/max_mem_allocated(gib)": 45.47,
1060
+ "step": 105
1061
+ }
1062
+ ],
1063
+ "logging_steps": 1,
1064
+ "max_steps": 315,
1065
+ "num_input_tokens_seen": 0,
1066
+ "num_train_epochs": 3,
1067
+ "save_steps": 500,
1068
+ "stateful_callbacks": {
1069
+ "TrainerControl": {
1070
+ "args": {
1071
+ "should_epoch_stop": false,
1072
+ "should_evaluate": false,
1073
+ "should_log": false,
1074
+ "should_save": true,
1075
+ "should_training_stop": false
1076
+ },
1077
+ "attributes": {}
1078
+ }
1079
+ },
1080
+ "total_flos": 1.6079273462857728e+16,
1081
+ "train_batch_size": 16,
1082
+ "trial_name": null,
1083
+ "trial_params": null
1084
+ }
checkpoint-105/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:232e77a355a770b00d79212785888b9356c3ed6653e1b633e2e2153e01eae951
3
+ size 6776
checkpoint-105/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-210/README.md ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B-Instruct-2507
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - axolotl
7
+ - base_model:adapter:Qwen/Qwen3-4B-Instruct-2507
8
+ - lora
9
+ - transformers
10
+ ---
11
+
12
+ # Model Card for Model ID
13
+
14
+ <!-- Provide a quick summary of what the model is/does. -->
15
+
16
+
17
+
18
+ ## Model Details
19
+
20
+ ### Model Description
21
+
22
+ <!-- Provide a longer summary of what this model is. -->
23
+
24
+
25
+
26
+ - **Developed by:** [More Information Needed]
27
+ - **Funded by [optional]:** [More Information Needed]
28
+ - **Shared by [optional]:** [More Information Needed]
29
+ - **Model type:** [More Information Needed]
30
+ - **Language(s) (NLP):** [More Information Needed]
31
+ - **License:** [More Information Needed]
32
+ - **Finetuned from model [optional]:** [More Information Needed]
33
+
34
+ ### Model Sources [optional]
35
+
36
+ <!-- Provide the basic links for the model. -->
37
+
38
+ - **Repository:** [More Information Needed]
39
+ - **Paper [optional]:** [More Information Needed]
40
+ - **Demo [optional]:** [More Information Needed]
41
+
42
+ ## Uses
43
+
44
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
45
+
46
+ ### Direct Use
47
+
48
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Downstream Use [optional]
53
+
54
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
55
+
56
+ [More Information Needed]
57
+
58
+ ### Out-of-Scope Use
59
+
60
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ## Bias, Risks, and Limitations
65
+
66
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
67
+
68
+ [More Information Needed]
69
+
70
+ ### Recommendations
71
+
72
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
73
+
74
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
75
+
76
+ ## How to Get Started with the Model
77
+
78
+ Use the code below to get started with the model.
79
+
80
+ [More Information Needed]
81
+
82
+ ## Training Details
83
+
84
+ ### Training Data
85
+
86
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
87
+
88
+ [More Information Needed]
89
+
90
+ ### Training Procedure
91
+
92
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
93
+
94
+ #### Preprocessing [optional]
95
+
96
+ [More Information Needed]
97
+
98
+
99
+ #### Training Hyperparameters
100
+
101
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
102
+
103
+ #### Speeds, Sizes, Times [optional]
104
+
105
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
106
+
107
+ [More Information Needed]
108
+
109
+ ## Evaluation
110
+
111
+ <!-- This section describes the evaluation protocols and provides the results. -->
112
+
113
+ ### Testing Data, Factors & Metrics
114
+
115
+ #### Testing Data
116
+
117
+ <!-- This should link to a Dataset Card if possible. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Factors
122
+
123
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
124
+
125
+ [More Information Needed]
126
+
127
+ #### Metrics
128
+
129
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
130
+
131
+ [More Information Needed]
132
+
133
+ ### Results
134
+
135
+ [More Information Needed]
136
+
137
+ #### Summary
138
+
139
+
140
+
141
+ ## Model Examination [optional]
142
+
143
+ <!-- Relevant interpretability work for the model goes here -->
144
+
145
+ [More Information Needed]
146
+
147
+ ## Environmental Impact
148
+
149
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
150
+
151
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
152
+
153
+ - **Hardware Type:** [More Information Needed]
154
+ - **Hours used:** [More Information Needed]
155
+ - **Cloud Provider:** [More Information Needed]
156
+ - **Compute Region:** [More Information Needed]
157
+ - **Carbon Emitted:** [More Information Needed]
158
+
159
+ ## Technical Specifications [optional]
160
+
161
+ ### Model Architecture and Objective
162
+
163
+ [More Information Needed]
164
+
165
+ ### Compute Infrastructure
166
+
167
+ [More Information Needed]
168
+
169
+ #### Hardware
170
+
171
+ [More Information Needed]
172
+
173
+ #### Software
174
+
175
+ [More Information Needed]
176
+
177
+ ## Citation [optional]
178
+
179
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
180
+
181
+ **BibTeX:**
182
+
183
+ [More Information Needed]
184
+
185
+ **APA:**
186
+
187
+ [More Information Needed]
188
+
189
+ ## Glossary [optional]
190
+
191
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
192
+
193
+ [More Information Needed]
194
+
195
+ ## More Information [optional]
196
+
197
+ [More Information Needed]
198
+
199
+ ## Model Card Authors [optional]
200
+
201
+ [More Information Needed]
202
+
203
+ ## Model Card Contact
204
+
205
+ [More Information Needed]
206
+ ### Framework versions
207
+
208
+ - PEFT 0.17.0
checkpoint-210/adapter_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen3-4B-Instruct-2507",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": null,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 8,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "q_proj",
29
+ "gate_proj",
30
+ "up_proj",
31
+ "v_proj",
32
+ "down_proj",
33
+ "k_proj",
34
+ "o_proj"
35
+ ],
36
+ "target_parameters": [],
37
+ "task_type": "CAUSAL_LM",
38
+ "trainable_token_indices": null,
39
+ "use_dora": false,
40
+ "use_qalora": false,
41
+ "use_rslora": false
42
+ }
checkpoint-210/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5ec294b0528c1151fe3a0a022417d82f5cf536ff45577bcf337eb630a64554c
3
+ size 66126768
checkpoint-210/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
checkpoint-210/chat_template.jinja ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if message.content is string %}
27
+ {%- set content = message.content %}
28
+ {%- else %}
29
+ {%- set content = '' %}
30
+ {%- endif %}
31
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
+ {%- elif message.role == "assistant" %}
34
+ {%- set reasoning_content = '' %}
35
+ {%- if message.reasoning_content is string %}
36
+ {%- set reasoning_content = message.reasoning_content %}
37
+ {%- else %}
38
+ {%- if '</think>' in content %}
39
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
+ {%- endif %}
42
+ {%- endif %}
43
+ {%- if loop.index0 > ns.last_query_index %}
44
+ {%- if loop.last or (not loop.last and reasoning_content) %}
45
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
+ {%- else %}
47
+ {{- '<|im_start|>' + message.role + '\n' + content }}
48
+ {%- endif %}
49
+ {%- else %}
50
+ {{- '<|im_start|>' + message.role + '\n' + content }}
51
+ {%- endif %}
52
+ {%- if message.tool_calls %}
53
+ {%- for tool_call in message.tool_calls %}
54
+ {%- if (loop.first and content) or (not loop.first) %}
55
+ {{- '\n' }}
56
+ {%- endif %}
57
+ {%- if tool_call.function %}
58
+ {%- set tool_call = tool_call.function %}
59
+ {%- endif %}
60
+ {{- '<tool_call>\n{"name": "' }}
61
+ {{- tool_call.name }}
62
+ {{- '", "arguments": ' }}
63
+ {%- if tool_call.arguments is string %}
64
+ {{- tool_call.arguments }}
65
+ {%- else %}
66
+ {{- tool_call.arguments | tojson }}
67
+ {%- endif %}
68
+ {{- '}\n</tool_call>' }}
69
+ {%- endfor %}
70
+ {%- endif %}
71
+ {{- '<|im_end|>\n' }}
72
+ {%- elif message.role == "tool" %}
73
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
74
+ {{- '<|im_start|>user' }}
75
+ {%- endif %}
76
+ {{- '\n<tool_response>\n' }}
77
+ {{- content }}
78
+ {{- '\n</tool_response>' }}
79
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
80
+ {{- '<|im_end|>\n' }}
81
+ {%- endif %}
82
+ {%- endif %}
83
+ {%- endfor %}
84
+ {%- if add_generation_prompt %}
85
+ {{- '<|im_start|>assistant\n' }}
86
+ {%- endif %}
checkpoint-210/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-210/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcad621f4ef1b1fae68830a27a7f62e353fb5835c7d2867f34698e414c59f356
3
+ size 34135860
checkpoint-210/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:476e7fc8020b17595f3ef4dad3e9bb8f476f39100a9c32966926557b9d34a5e3
3
+ size 14512
checkpoint-210/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2be463597d503ef315c7bcad8c17a62b22c0357b52eab8a909a62ea745c2e290
3
+ size 14512
checkpoint-210/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b3262c00bc04620eb4ac6df142b7b0caf7dc06645441af56eb661f2b3afbc3f
3
+ size 1064
checkpoint-210/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
checkpoint-210/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
checkpoint-210/tokenizer_config.json ADDED
@@ -0,0 +1,239 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|im_end|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 262144,
235
+ "pad_token": "<|endoftext|>",
236
+ "split_special_tokens": false,
237
+ "tokenizer_class": "Qwen2Tokenizer",
238
+ "unk_token": null
239
+ }
checkpoint-210/trainer_state.json ADDED
@@ -0,0 +1,2134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 210,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.009523809523809525,
14
+ "grad_norm": 2.447430372238159,
15
+ "learning_rate": 0.0,
16
+ "loss": 1.8171,
17
+ "memory/device_mem_reserved(gib)": 36.42,
18
+ "memory/max_mem_active(gib)": 35.17,
19
+ "memory/max_mem_allocated(gib)": 35.17,
20
+ "step": 1
21
+ },
22
+ {
23
+ "epoch": 0.01904761904761905,
24
+ "grad_norm": 2.280487537384033,
25
+ "learning_rate": 2.2222222222222223e-05,
26
+ "loss": 1.7544,
27
+ "memory/device_mem_reserved(gib)": 36.45,
28
+ "memory/max_mem_active(gib)": 35.23,
29
+ "memory/max_mem_allocated(gib)": 35.23,
30
+ "step": 2
31
+ },
32
+ {
33
+ "epoch": 0.02857142857142857,
34
+ "grad_norm": 2.248966693878174,
35
+ "learning_rate": 4.4444444444444447e-05,
36
+ "loss": 1.7546,
37
+ "memory/device_mem_reserved(gib)": 47.1,
38
+ "memory/max_mem_active(gib)": 45.45,
39
+ "memory/max_mem_allocated(gib)": 45.45,
40
+ "step": 3
41
+ },
42
+ {
43
+ "epoch": 0.0380952380952381,
44
+ "grad_norm": 2.2559540271759033,
45
+ "learning_rate": 6.666666666666667e-05,
46
+ "loss": 1.8011,
47
+ "memory/device_mem_reserved(gib)": 47.1,
48
+ "memory/max_mem_active(gib)": 45.45,
49
+ "memory/max_mem_allocated(gib)": 45.45,
50
+ "step": 4
51
+ },
52
+ {
53
+ "epoch": 0.047619047619047616,
54
+ "grad_norm": 2.149872064590454,
55
+ "learning_rate": 8.888888888888889e-05,
56
+ "loss": 1.6465,
57
+ "memory/device_mem_reserved(gib)": 47.1,
58
+ "memory/max_mem_active(gib)": 45.45,
59
+ "memory/max_mem_allocated(gib)": 45.45,
60
+ "step": 5
61
+ },
62
+ {
63
+ "epoch": 0.05714285714285714,
64
+ "grad_norm": 1.6111228466033936,
65
+ "learning_rate": 0.00011111111111111112,
66
+ "loss": 1.5935,
67
+ "memory/device_mem_reserved(gib)": 47.1,
68
+ "memory/max_mem_active(gib)": 45.45,
69
+ "memory/max_mem_allocated(gib)": 45.45,
70
+ "step": 6
71
+ },
72
+ {
73
+ "epoch": 0.06666666666666667,
74
+ "grad_norm": 1.363952875137329,
75
+ "learning_rate": 0.00013333333333333334,
76
+ "loss": 1.4371,
77
+ "memory/device_mem_reserved(gib)": 47.1,
78
+ "memory/max_mem_active(gib)": 45.45,
79
+ "memory/max_mem_allocated(gib)": 45.45,
80
+ "step": 7
81
+ },
82
+ {
83
+ "epoch": 0.0761904761904762,
84
+ "grad_norm": 1.1439425945281982,
85
+ "learning_rate": 0.00015555555555555556,
86
+ "loss": 1.2409,
87
+ "memory/device_mem_reserved(gib)": 47.1,
88
+ "memory/max_mem_active(gib)": 45.45,
89
+ "memory/max_mem_allocated(gib)": 45.45,
90
+ "step": 8
91
+ },
92
+ {
93
+ "epoch": 0.08571428571428572,
94
+ "grad_norm": 0.877252995967865,
95
+ "learning_rate": 0.00017777777777777779,
96
+ "loss": 1.1083,
97
+ "memory/device_mem_reserved(gib)": 47.1,
98
+ "memory/max_mem_active(gib)": 45.45,
99
+ "memory/max_mem_allocated(gib)": 45.45,
100
+ "step": 9
101
+ },
102
+ {
103
+ "epoch": 0.09523809523809523,
104
+ "grad_norm": 0.8071479797363281,
105
+ "learning_rate": 0.0002,
106
+ "loss": 0.9937,
107
+ "memory/device_mem_reserved(gib)": 47.1,
108
+ "memory/max_mem_active(gib)": 45.45,
109
+ "memory/max_mem_allocated(gib)": 45.45,
110
+ "step": 10
111
+ },
112
+ {
113
+ "epoch": 0.10476190476190476,
114
+ "grad_norm": 0.7772606015205383,
115
+ "learning_rate": 0.00019999472984871732,
116
+ "loss": 0.9235,
117
+ "memory/device_mem_reserved(gib)": 47.1,
118
+ "memory/max_mem_active(gib)": 45.45,
119
+ "memory/max_mem_allocated(gib)": 45.45,
120
+ "step": 11
121
+ },
122
+ {
123
+ "epoch": 0.11428571428571428,
124
+ "grad_norm": 0.811764657497406,
125
+ "learning_rate": 0.00019997891995035912,
126
+ "loss": 0.7635,
127
+ "memory/device_mem_reserved(gib)": 47.1,
128
+ "memory/max_mem_active(gib)": 45.45,
129
+ "memory/max_mem_allocated(gib)": 45.45,
130
+ "step": 12
131
+ },
132
+ {
133
+ "epoch": 0.12380952380952381,
134
+ "grad_norm": 0.8423659801483154,
135
+ "learning_rate": 0.0001999525719713366,
136
+ "loss": 0.6836,
137
+ "memory/device_mem_reserved(gib)": 47.1,
138
+ "memory/max_mem_active(gib)": 45.45,
139
+ "memory/max_mem_allocated(gib)": 45.45,
140
+ "step": 13
141
+ },
142
+ {
143
+ "epoch": 0.13333333333333333,
144
+ "grad_norm": 0.7858280539512634,
145
+ "learning_rate": 0.0001999156886888064,
146
+ "loss": 0.6159,
147
+ "memory/device_mem_reserved(gib)": 47.1,
148
+ "memory/max_mem_active(gib)": 45.45,
149
+ "memory/max_mem_allocated(gib)": 45.45,
150
+ "step": 14
151
+ },
152
+ {
153
+ "epoch": 0.14285714285714285,
154
+ "grad_norm": 0.739560604095459,
155
+ "learning_rate": 0.00019986827399037812,
156
+ "loss": 0.5082,
157
+ "memory/device_mem_reserved(gib)": 47.1,
158
+ "memory/max_mem_active(gib)": 45.45,
159
+ "memory/max_mem_allocated(gib)": 45.45,
160
+ "step": 15
161
+ },
162
+ {
163
+ "epoch": 0.1523809523809524,
164
+ "grad_norm": 0.6787500381469727,
165
+ "learning_rate": 0.00019981033287370443,
166
+ "loss": 0.4553,
167
+ "memory/device_mem_reserved(gib)": 47.1,
168
+ "memory/max_mem_active(gib)": 45.45,
169
+ "memory/max_mem_allocated(gib)": 45.45,
170
+ "step": 16
171
+ },
172
+ {
173
+ "epoch": 0.1619047619047619,
174
+ "grad_norm": 0.6258607506752014,
175
+ "learning_rate": 0.00019974187144595432,
176
+ "loss": 0.3913,
177
+ "memory/device_mem_reserved(gib)": 47.1,
178
+ "memory/max_mem_active(gib)": 45.45,
179
+ "memory/max_mem_allocated(gib)": 45.45,
180
+ "step": 17
181
+ },
182
+ {
183
+ "epoch": 0.17142857142857143,
184
+ "grad_norm": 0.47939175367355347,
185
+ "learning_rate": 0.00019966289692316944,
186
+ "loss": 0.3048,
187
+ "memory/device_mem_reserved(gib)": 47.1,
188
+ "memory/max_mem_active(gib)": 45.45,
189
+ "memory/max_mem_allocated(gib)": 45.45,
190
+ "step": 18
191
+ },
192
+ {
193
+ "epoch": 0.18095238095238095,
194
+ "grad_norm": 0.49321815371513367,
195
+ "learning_rate": 0.00019957341762950344,
196
+ "loss": 0.3417,
197
+ "memory/device_mem_reserved(gib)": 47.1,
198
+ "memory/max_mem_active(gib)": 45.45,
199
+ "memory/max_mem_allocated(gib)": 45.45,
200
+ "step": 19
201
+ },
202
+ {
203
+ "epoch": 0.19047619047619047,
204
+ "grad_norm": 0.4252234101295471,
205
+ "learning_rate": 0.00019947344299634464,
206
+ "loss": 0.3033,
207
+ "memory/device_mem_reserved(gib)": 47.1,
208
+ "memory/max_mem_active(gib)": 45.45,
209
+ "memory/max_mem_allocated(gib)": 45.45,
210
+ "step": 20
211
+ },
212
+ {
213
+ "epoch": 0.2,
214
+ "grad_norm": 0.46225014328956604,
215
+ "learning_rate": 0.00019936298356132176,
216
+ "loss": 0.2984,
217
+ "memory/device_mem_reserved(gib)": 47.1,
218
+ "memory/max_mem_active(gib)": 45.45,
219
+ "memory/max_mem_allocated(gib)": 45.45,
220
+ "step": 21
221
+ },
222
+ {
223
+ "epoch": 0.20952380952380953,
224
+ "grad_norm": 0.4106466472148895,
225
+ "learning_rate": 0.0001992420509671936,
226
+ "loss": 0.2764,
227
+ "memory/device_mem_reserved(gib)": 47.12,
228
+ "memory/max_mem_active(gib)": 45.46,
229
+ "memory/max_mem_allocated(gib)": 45.46,
230
+ "step": 22
231
+ },
232
+ {
233
+ "epoch": 0.21904761904761905,
234
+ "grad_norm": 0.3764067590236664,
235
+ "learning_rate": 0.00019911065796062135,
236
+ "loss": 0.2352,
237
+ "memory/device_mem_reserved(gib)": 47.12,
238
+ "memory/max_mem_active(gib)": 45.46,
239
+ "memory/max_mem_allocated(gib)": 45.46,
240
+ "step": 23
241
+ },
242
+ {
243
+ "epoch": 0.22857142857142856,
244
+ "grad_norm": 0.38518962264060974,
245
+ "learning_rate": 0.00019896881839082556,
246
+ "loss": 0.2239,
247
+ "memory/device_mem_reserved(gib)": 47.12,
248
+ "memory/max_mem_active(gib)": 45.46,
249
+ "memory/max_mem_allocated(gib)": 45.46,
250
+ "step": 24
251
+ },
252
+ {
253
+ "epoch": 0.23809523809523808,
254
+ "grad_norm": 0.39728403091430664,
255
+ "learning_rate": 0.00019881654720812594,
256
+ "loss": 0.2192,
257
+ "memory/device_mem_reserved(gib)": 47.13,
258
+ "memory/max_mem_active(gib)": 45.46,
259
+ "memory/max_mem_allocated(gib)": 45.46,
260
+ "step": 25
261
+ },
262
+ {
263
+ "epoch": 0.24761904761904763,
264
+ "grad_norm": 0.3667093813419342,
265
+ "learning_rate": 0.00019865386046236596,
266
+ "loss": 0.2834,
267
+ "memory/device_mem_reserved(gib)": 47.13,
268
+ "memory/max_mem_active(gib)": 45.46,
269
+ "memory/max_mem_allocated(gib)": 45.46,
270
+ "step": 26
271
+ },
272
+ {
273
+ "epoch": 0.2571428571428571,
274
+ "grad_norm": 0.32502660155296326,
275
+ "learning_rate": 0.00019848077530122083,
276
+ "loss": 0.2166,
277
+ "memory/device_mem_reserved(gib)": 47.13,
278
+ "memory/max_mem_active(gib)": 45.46,
279
+ "memory/max_mem_allocated(gib)": 45.46,
280
+ "step": 27
281
+ },
282
+ {
283
+ "epoch": 0.26666666666666666,
284
+ "grad_norm": 0.31225672364234924,
285
+ "learning_rate": 0.0001982973099683902,
286
+ "loss": 0.2238,
287
+ "memory/device_mem_reserved(gib)": 47.13,
288
+ "memory/max_mem_active(gib)": 45.46,
289
+ "memory/max_mem_allocated(gib)": 45.46,
290
+ "step": 28
291
+ },
292
+ {
293
+ "epoch": 0.2761904761904762,
294
+ "grad_norm": 0.27448731660842896,
295
+ "learning_rate": 0.00019810348380167527,
296
+ "loss": 0.1983,
297
+ "memory/device_mem_reserved(gib)": 47.13,
298
+ "memory/max_mem_active(gib)": 45.46,
299
+ "memory/max_mem_allocated(gib)": 45.46,
300
+ "step": 29
301
+ },
302
+ {
303
+ "epoch": 0.2857142857142857,
304
+ "grad_norm": 0.32014158368110657,
305
+ "learning_rate": 0.00019789931723094046,
306
+ "loss": 0.2554,
307
+ "memory/device_mem_reserved(gib)": 47.13,
308
+ "memory/max_mem_active(gib)": 45.46,
309
+ "memory/max_mem_allocated(gib)": 45.46,
310
+ "step": 30
311
+ },
312
+ {
313
+ "epoch": 0.29523809523809524,
314
+ "grad_norm": 0.39480143785476685,
315
+ "learning_rate": 0.0001976848317759601,
316
+ "loss": 0.2438,
317
+ "memory/device_mem_reserved(gib)": 47.13,
318
+ "memory/max_mem_active(gib)": 45.46,
319
+ "memory/max_mem_allocated(gib)": 45.46,
320
+ "step": 31
321
+ },
322
+ {
323
+ "epoch": 0.3047619047619048,
324
+ "grad_norm": 0.28004321455955505,
325
+ "learning_rate": 0.00019746005004415005,
326
+ "loss": 0.2145,
327
+ "memory/device_mem_reserved(gib)": 47.13,
328
+ "memory/max_mem_active(gib)": 45.46,
329
+ "memory/max_mem_allocated(gib)": 45.46,
330
+ "step": 32
331
+ },
332
+ {
333
+ "epoch": 0.3142857142857143,
334
+ "grad_norm": 0.2754361033439636,
335
+ "learning_rate": 0.00019722499572818496,
336
+ "loss": 0.2284,
337
+ "memory/device_mem_reserved(gib)": 47.13,
338
+ "memory/max_mem_active(gib)": 45.46,
339
+ "memory/max_mem_allocated(gib)": 45.46,
340
+ "step": 33
341
+ },
342
+ {
343
+ "epoch": 0.3238095238095238,
344
+ "grad_norm": 0.3160915970802307,
345
+ "learning_rate": 0.00019697969360350098,
346
+ "loss": 0.2482,
347
+ "memory/device_mem_reserved(gib)": 47.13,
348
+ "memory/max_mem_active(gib)": 45.47,
349
+ "memory/max_mem_allocated(gib)": 45.47,
350
+ "step": 34
351
+ },
352
+ {
353
+ "epoch": 0.3333333333333333,
354
+ "grad_norm": 0.24699093401432037,
355
+ "learning_rate": 0.00019672416952568416,
356
+ "loss": 0.1916,
357
+ "memory/device_mem_reserved(gib)": 47.13,
358
+ "memory/max_mem_active(gib)": 45.47,
359
+ "memory/max_mem_allocated(gib)": 45.47,
360
+ "step": 35
361
+ },
362
+ {
363
+ "epoch": 0.34285714285714286,
364
+ "grad_norm": 0.2674485743045807,
365
+ "learning_rate": 0.00019645845042774553,
366
+ "loss": 0.2204,
367
+ "memory/device_mem_reserved(gib)": 47.13,
368
+ "memory/max_mem_active(gib)": 45.47,
369
+ "memory/max_mem_allocated(gib)": 45.47,
370
+ "step": 36
371
+ },
372
+ {
373
+ "epoch": 0.3523809523809524,
374
+ "grad_norm": 0.33608749508857727,
375
+ "learning_rate": 0.00019618256431728194,
376
+ "loss": 0.235,
377
+ "memory/device_mem_reserved(gib)": 47.13,
378
+ "memory/max_mem_active(gib)": 45.47,
379
+ "memory/max_mem_allocated(gib)": 45.47,
380
+ "step": 37
381
+ },
382
+ {
383
+ "epoch": 0.3619047619047619,
384
+ "grad_norm": 0.29581594467163086,
385
+ "learning_rate": 0.00019589654027352414,
386
+ "loss": 0.2175,
387
+ "memory/device_mem_reserved(gib)": 47.13,
388
+ "memory/max_mem_active(gib)": 45.47,
389
+ "memory/max_mem_allocated(gib)": 45.47,
390
+ "step": 38
391
+ },
392
+ {
393
+ "epoch": 0.37142857142857144,
394
+ "grad_norm": 0.3526640832424164,
395
+ "learning_rate": 0.0001956004084442718,
396
+ "loss": 0.1989,
397
+ "memory/device_mem_reserved(gib)": 47.13,
398
+ "memory/max_mem_active(gib)": 45.47,
399
+ "memory/max_mem_allocated(gib)": 45.47,
400
+ "step": 39
401
+ },
402
+ {
403
+ "epoch": 0.38095238095238093,
404
+ "grad_norm": 0.2882135808467865,
405
+ "learning_rate": 0.00019529420004271567,
406
+ "loss": 0.2144,
407
+ "memory/device_mem_reserved(gib)": 47.13,
408
+ "memory/max_mem_active(gib)": 45.47,
409
+ "memory/max_mem_allocated(gib)": 45.47,
410
+ "step": 40
411
+ },
412
+ {
413
+ "epoch": 0.3904761904761905,
414
+ "grad_norm": 0.2675183117389679,
415
+ "learning_rate": 0.0001949779473441478,
416
+ "loss": 0.1987,
417
+ "memory/device_mem_reserved(gib)": 47.13,
418
+ "memory/max_mem_active(gib)": 45.47,
419
+ "memory/max_mem_allocated(gib)": 45.47,
420
+ "step": 41
421
+ },
422
+ {
423
+ "epoch": 0.4,
424
+ "grad_norm": 0.2908264696598053,
425
+ "learning_rate": 0.00019465168368255946,
426
+ "loss": 0.1893,
427
+ "memory/device_mem_reserved(gib)": 47.13,
428
+ "memory/max_mem_active(gib)": 45.47,
429
+ "memory/max_mem_allocated(gib)": 45.47,
430
+ "step": 42
431
+ },
432
+ {
433
+ "epoch": 0.4095238095238095,
434
+ "grad_norm": 0.3211234211921692,
435
+ "learning_rate": 0.00019431544344712776,
436
+ "loss": 0.2008,
437
+ "memory/device_mem_reserved(gib)": 47.13,
438
+ "memory/max_mem_active(gib)": 45.47,
439
+ "memory/max_mem_allocated(gib)": 45.47,
440
+ "step": 43
441
+ },
442
+ {
443
+ "epoch": 0.41904761904761906,
444
+ "grad_norm": 0.2795293927192688,
445
+ "learning_rate": 0.00019396926207859084,
446
+ "loss": 0.1935,
447
+ "memory/device_mem_reserved(gib)": 47.13,
448
+ "memory/max_mem_active(gib)": 45.47,
449
+ "memory/max_mem_allocated(gib)": 45.47,
450
+ "step": 44
451
+ },
452
+ {
453
+ "epoch": 0.42857142857142855,
454
+ "grad_norm": 0.28679871559143066,
455
+ "learning_rate": 0.00019361317606551238,
456
+ "loss": 0.1987,
457
+ "memory/device_mem_reserved(gib)": 47.13,
458
+ "memory/max_mem_active(gib)": 45.47,
459
+ "memory/max_mem_allocated(gib)": 45.47,
460
+ "step": 45
461
+ },
462
+ {
463
+ "epoch": 0.4380952380952381,
464
+ "grad_norm": 0.3165067732334137,
465
+ "learning_rate": 0.00019324722294043558,
466
+ "loss": 0.2118,
467
+ "memory/device_mem_reserved(gib)": 47.13,
468
+ "memory/max_mem_active(gib)": 45.47,
469
+ "memory/max_mem_allocated(gib)": 45.47,
470
+ "step": 46
471
+ },
472
+ {
473
+ "epoch": 0.44761904761904764,
474
+ "grad_norm": 0.25720369815826416,
475
+ "learning_rate": 0.00019287144127592704,
476
+ "loss": 0.1962,
477
+ "memory/device_mem_reserved(gib)": 47.13,
478
+ "memory/max_mem_active(gib)": 45.47,
479
+ "memory/max_mem_allocated(gib)": 45.47,
480
+ "step": 47
481
+ },
482
+ {
483
+ "epoch": 0.45714285714285713,
484
+ "grad_norm": 0.2577202022075653,
485
+ "learning_rate": 0.0001924858706805112,
486
+ "loss": 0.2022,
487
+ "memory/device_mem_reserved(gib)": 47.13,
488
+ "memory/max_mem_active(gib)": 45.47,
489
+ "memory/max_mem_allocated(gib)": 45.47,
490
+ "step": 48
491
+ },
492
+ {
493
+ "epoch": 0.4666666666666667,
494
+ "grad_norm": 0.2605520188808441,
495
+ "learning_rate": 0.0001920905517944954,
496
+ "loss": 0.1891,
497
+ "memory/device_mem_reserved(gib)": 47.13,
498
+ "memory/max_mem_active(gib)": 45.47,
499
+ "memory/max_mem_allocated(gib)": 45.47,
500
+ "step": 49
501
+ },
502
+ {
503
+ "epoch": 0.47619047619047616,
504
+ "grad_norm": 0.2890350818634033,
505
+ "learning_rate": 0.00019168552628568631,
506
+ "loss": 0.17,
507
+ "memory/device_mem_reserved(gib)": 47.13,
508
+ "memory/max_mem_active(gib)": 45.47,
509
+ "memory/max_mem_allocated(gib)": 45.47,
510
+ "step": 50
511
+ },
512
+ {
513
+ "epoch": 0.4857142857142857,
514
+ "grad_norm": 0.25532475113868713,
515
+ "learning_rate": 0.00019127083684499806,
516
+ "loss": 0.1796,
517
+ "memory/device_mem_reserved(gib)": 47.13,
518
+ "memory/max_mem_active(gib)": 45.47,
519
+ "memory/max_mem_allocated(gib)": 45.47,
520
+ "step": 51
521
+ },
522
+ {
523
+ "epoch": 0.49523809523809526,
524
+ "grad_norm": 0.2963904142379761,
525
+ "learning_rate": 0.00019084652718195238,
526
+ "loss": 0.19,
527
+ "memory/device_mem_reserved(gib)": 47.13,
528
+ "memory/max_mem_active(gib)": 45.47,
529
+ "memory/max_mem_allocated(gib)": 45.47,
530
+ "step": 52
531
+ },
532
+ {
533
+ "epoch": 0.5047619047619047,
534
+ "grad_norm": 0.24455095827579498,
535
+ "learning_rate": 0.0001904126420200716,
536
+ "loss": 0.2017,
537
+ "memory/device_mem_reserved(gib)": 47.13,
538
+ "memory/max_mem_active(gib)": 45.47,
539
+ "memory/max_mem_allocated(gib)": 45.47,
540
+ "step": 53
541
+ },
542
+ {
543
+ "epoch": 0.5142857142857142,
544
+ "grad_norm": 0.2518145442008972,
545
+ "learning_rate": 0.00018996922709216455,
546
+ "loss": 0.1896,
547
+ "memory/device_mem_reserved(gib)": 47.13,
548
+ "memory/max_mem_active(gib)": 45.47,
549
+ "memory/max_mem_allocated(gib)": 45.47,
550
+ "step": 54
551
+ },
552
+ {
553
+ "epoch": 0.5238095238095238,
554
+ "grad_norm": 0.26225805282592773,
555
+ "learning_rate": 0.00018951632913550626,
556
+ "loss": 0.1893,
557
+ "memory/device_mem_reserved(gib)": 47.13,
558
+ "memory/max_mem_active(gib)": 45.47,
559
+ "memory/max_mem_allocated(gib)": 45.47,
560
+ "step": 55
561
+ },
562
+ {
563
+ "epoch": 0.5333333333333333,
564
+ "grad_norm": 0.25169169902801514,
565
+ "learning_rate": 0.00018905399588691163,
566
+ "loss": 0.1939,
567
+ "memory/device_mem_reserved(gib)": 47.13,
568
+ "memory/max_mem_active(gib)": 45.47,
569
+ "memory/max_mem_allocated(gib)": 45.47,
570
+ "step": 56
571
+ },
572
+ {
573
+ "epoch": 0.5428571428571428,
574
+ "grad_norm": 0.23847182095050812,
575
+ "learning_rate": 0.00018858227607770398,
576
+ "loss": 0.1686,
577
+ "memory/device_mem_reserved(gib)": 47.13,
578
+ "memory/max_mem_active(gib)": 45.47,
579
+ "memory/max_mem_allocated(gib)": 45.47,
580
+ "step": 57
581
+ },
582
+ {
583
+ "epoch": 0.5523809523809524,
584
+ "grad_norm": 0.30888280272483826,
585
+ "learning_rate": 0.00018810121942857845,
586
+ "loss": 0.1768,
587
+ "memory/device_mem_reserved(gib)": 47.13,
588
+ "memory/max_mem_active(gib)": 45.47,
589
+ "memory/max_mem_allocated(gib)": 45.47,
590
+ "step": 58
591
+ },
592
+ {
593
+ "epoch": 0.5619047619047619,
594
+ "grad_norm": 0.27275580167770386,
595
+ "learning_rate": 0.00018761087664436138,
596
+ "loss": 0.1941,
597
+ "memory/device_mem_reserved(gib)": 47.13,
598
+ "memory/max_mem_active(gib)": 45.47,
599
+ "memory/max_mem_allocated(gib)": 45.47,
600
+ "step": 59
601
+ },
602
+ {
603
+ "epoch": 0.5714285714285714,
604
+ "grad_norm": 0.22851359844207764,
605
+ "learning_rate": 0.00018711129940866575,
606
+ "loss": 0.1658,
607
+ "memory/device_mem_reserved(gib)": 47.13,
608
+ "memory/max_mem_active(gib)": 45.47,
609
+ "memory/max_mem_allocated(gib)": 45.47,
610
+ "step": 60
611
+ },
612
+ {
613
+ "epoch": 0.580952380952381,
614
+ "grad_norm": 0.25515425205230713,
615
+ "learning_rate": 0.00018660254037844388,
616
+ "loss": 0.1858,
617
+ "memory/device_mem_reserved(gib)": 47.13,
618
+ "memory/max_mem_active(gib)": 45.47,
619
+ "memory/max_mem_allocated(gib)": 45.47,
620
+ "step": 61
621
+ },
622
+ {
623
+ "epoch": 0.5904761904761905,
624
+ "grad_norm": 0.24301765859127045,
625
+ "learning_rate": 0.00018608465317843678,
626
+ "loss": 0.1892,
627
+ "memory/device_mem_reserved(gib)": 47.13,
628
+ "memory/max_mem_active(gib)": 45.47,
629
+ "memory/max_mem_allocated(gib)": 45.47,
630
+ "step": 62
631
+ },
632
+ {
633
+ "epoch": 0.6,
634
+ "grad_norm": 0.24329794943332672,
635
+ "learning_rate": 0.00018555769239552233,
636
+ "loss": 0.1719,
637
+ "memory/device_mem_reserved(gib)": 47.13,
638
+ "memory/max_mem_active(gib)": 45.47,
639
+ "memory/max_mem_allocated(gib)": 45.47,
640
+ "step": 63
641
+ },
642
+ {
643
+ "epoch": 0.6095238095238096,
644
+ "grad_norm": 0.3206610381603241,
645
+ "learning_rate": 0.00018502171357296144,
646
+ "loss": 0.2075,
647
+ "memory/device_mem_reserved(gib)": 47.13,
648
+ "memory/max_mem_active(gib)": 45.47,
649
+ "memory/max_mem_allocated(gib)": 45.47,
650
+ "step": 64
651
+ },
652
+ {
653
+ "epoch": 0.6190476190476191,
654
+ "grad_norm": 0.3089618980884552,
655
+ "learning_rate": 0.00018447677320454367,
656
+ "loss": 0.185,
657
+ "memory/device_mem_reserved(gib)": 47.13,
658
+ "memory/max_mem_active(gib)": 45.47,
659
+ "memory/max_mem_allocated(gib)": 45.47,
660
+ "step": 65
661
+ },
662
+ {
663
+ "epoch": 0.6285714285714286,
664
+ "grad_norm": 0.291916161775589,
665
+ "learning_rate": 0.00018392292872863267,
666
+ "loss": 0.1739,
667
+ "memory/device_mem_reserved(gib)": 47.13,
668
+ "memory/max_mem_active(gib)": 45.47,
669
+ "memory/max_mem_allocated(gib)": 45.47,
670
+ "step": 66
671
+ },
672
+ {
673
+ "epoch": 0.638095238095238,
674
+ "grad_norm": 0.25783228874206543,
675
+ "learning_rate": 0.00018336023852211195,
676
+ "loss": 0.2061,
677
+ "memory/device_mem_reserved(gib)": 47.13,
678
+ "memory/max_mem_active(gib)": 45.47,
679
+ "memory/max_mem_allocated(gib)": 45.47,
680
+ "step": 67
681
+ },
682
+ {
683
+ "epoch": 0.6476190476190476,
684
+ "grad_norm": 0.310614675283432,
685
+ "learning_rate": 0.00018278876189423179,
686
+ "loss": 0.1984,
687
+ "memory/device_mem_reserved(gib)": 47.13,
688
+ "memory/max_mem_active(gib)": 45.47,
689
+ "memory/max_mem_allocated(gib)": 45.47,
690
+ "step": 68
691
+ },
692
+ {
693
+ "epoch": 0.6571428571428571,
694
+ "grad_norm": 0.23836325109004974,
695
+ "learning_rate": 0.00018220855908035783,
696
+ "loss": 0.1742,
697
+ "memory/device_mem_reserved(gib)": 47.13,
698
+ "memory/max_mem_active(gib)": 45.47,
699
+ "memory/max_mem_allocated(gib)": 45.47,
700
+ "step": 69
701
+ },
702
+ {
703
+ "epoch": 0.6666666666666666,
704
+ "grad_norm": 0.30577945709228516,
705
+ "learning_rate": 0.0001816196912356222,
706
+ "loss": 0.1954,
707
+ "memory/device_mem_reserved(gib)": 47.13,
708
+ "memory/max_mem_active(gib)": 45.47,
709
+ "memory/max_mem_allocated(gib)": 45.47,
710
+ "step": 70
711
+ },
712
+ {
713
+ "epoch": 0.6761904761904762,
714
+ "grad_norm": 0.3177741467952728,
715
+ "learning_rate": 0.00018102222042847737,
716
+ "loss": 0.193,
717
+ "memory/device_mem_reserved(gib)": 47.13,
718
+ "memory/max_mem_active(gib)": 45.47,
719
+ "memory/max_mem_allocated(gib)": 45.47,
720
+ "step": 71
721
+ },
722
+ {
723
+ "epoch": 0.6857142857142857,
724
+ "grad_norm": 0.3444816470146179,
725
+ "learning_rate": 0.00018041620963415417,
726
+ "loss": 0.1837,
727
+ "memory/device_mem_reserved(gib)": 47.13,
728
+ "memory/max_mem_active(gib)": 45.47,
729
+ "memory/max_mem_allocated(gib)": 45.47,
730
+ "step": 72
731
+ },
732
+ {
733
+ "epoch": 0.6952380952380952,
734
+ "grad_norm": 0.3139638900756836,
735
+ "learning_rate": 0.000179801722728024,
736
+ "loss": 0.1902,
737
+ "memory/device_mem_reserved(gib)": 47.13,
738
+ "memory/max_mem_active(gib)": 45.47,
739
+ "memory/max_mem_allocated(gib)": 45.47,
740
+ "step": 73
741
+ },
742
+ {
743
+ "epoch": 0.7047619047619048,
744
+ "grad_norm": 0.24754184484481812,
745
+ "learning_rate": 0.00017917882447886582,
746
+ "loss": 0.1612,
747
+ "memory/device_mem_reserved(gib)": 47.13,
748
+ "memory/max_mem_active(gib)": 45.47,
749
+ "memory/max_mem_allocated(gib)": 45.47,
750
+ "step": 74
751
+ },
752
+ {
753
+ "epoch": 0.7142857142857143,
754
+ "grad_norm": 0.23204036056995392,
755
+ "learning_rate": 0.00017854758054203988,
756
+ "loss": 0.1694,
757
+ "memory/device_mem_reserved(gib)": 47.13,
758
+ "memory/max_mem_active(gib)": 45.47,
759
+ "memory/max_mem_allocated(gib)": 45.47,
760
+ "step": 75
761
+ },
762
+ {
763
+ "epoch": 0.7238095238095238,
764
+ "grad_norm": 0.2823260426521301,
765
+ "learning_rate": 0.00017790805745256704,
766
+ "loss": 0.1739,
767
+ "memory/device_mem_reserved(gib)": 47.13,
768
+ "memory/max_mem_active(gib)": 45.47,
769
+ "memory/max_mem_allocated(gib)": 45.47,
770
+ "step": 76
771
+ },
772
+ {
773
+ "epoch": 0.7333333333333333,
774
+ "grad_norm": 0.3193572163581848,
775
+ "learning_rate": 0.0001772603226181159,
776
+ "loss": 0.1924,
777
+ "memory/device_mem_reserved(gib)": 47.13,
778
+ "memory/max_mem_active(gib)": 45.47,
779
+ "memory/max_mem_allocated(gib)": 45.47,
780
+ "step": 77
781
+ },
782
+ {
783
+ "epoch": 0.7428571428571429,
784
+ "grad_norm": 0.30530717968940735,
785
+ "learning_rate": 0.0001766044443118978,
786
+ "loss": 0.1375,
787
+ "memory/device_mem_reserved(gib)": 47.13,
788
+ "memory/max_mem_active(gib)": 45.47,
789
+ "memory/max_mem_allocated(gib)": 45.47,
790
+ "step": 78
791
+ },
792
+ {
793
+ "epoch": 0.7523809523809524,
794
+ "grad_norm": 0.2974378764629364,
795
+ "learning_rate": 0.00017594049166547073,
796
+ "loss": 0.1866,
797
+ "memory/device_mem_reserved(gib)": 47.13,
798
+ "memory/max_mem_active(gib)": 45.47,
799
+ "memory/max_mem_allocated(gib)": 45.47,
800
+ "step": 79
801
+ },
802
+ {
803
+ "epoch": 0.7619047619047619,
804
+ "grad_norm": 0.26553574204444885,
805
+ "learning_rate": 0.00017526853466145244,
806
+ "loss": 0.193,
807
+ "memory/device_mem_reserved(gib)": 47.13,
808
+ "memory/max_mem_active(gib)": 45.47,
809
+ "memory/max_mem_allocated(gib)": 45.47,
810
+ "step": 80
811
+ },
812
+ {
813
+ "epoch": 0.7714285714285715,
814
+ "grad_norm": 0.32059594988822937,
815
+ "learning_rate": 0.00017458864412614434,
816
+ "loss": 0.1822,
817
+ "memory/device_mem_reserved(gib)": 47.13,
818
+ "memory/max_mem_active(gib)": 45.47,
819
+ "memory/max_mem_allocated(gib)": 45.47,
820
+ "step": 81
821
+ },
822
+ {
823
+ "epoch": 0.780952380952381,
824
+ "grad_norm": 0.3260822594165802,
825
+ "learning_rate": 0.00017390089172206592,
826
+ "loss": 0.1734,
827
+ "memory/device_mem_reserved(gib)": 47.13,
828
+ "memory/max_mem_active(gib)": 45.47,
829
+ "memory/max_mem_allocated(gib)": 45.47,
830
+ "step": 82
831
+ },
832
+ {
833
+ "epoch": 0.7904761904761904,
834
+ "grad_norm": 0.2754499912261963,
835
+ "learning_rate": 0.00017320534994040148,
836
+ "loss": 0.1814,
837
+ "memory/device_mem_reserved(gib)": 47.13,
838
+ "memory/max_mem_active(gib)": 45.47,
839
+ "memory/max_mem_allocated(gib)": 45.47,
840
+ "step": 83
841
+ },
842
+ {
843
+ "epoch": 0.8,
844
+ "grad_norm": 0.2639872431755066,
845
+ "learning_rate": 0.00017250209209335927,
846
+ "loss": 0.1764,
847
+ "memory/device_mem_reserved(gib)": 47.13,
848
+ "memory/max_mem_active(gib)": 45.47,
849
+ "memory/max_mem_allocated(gib)": 45.47,
850
+ "step": 84
851
+ },
852
+ {
853
+ "epoch": 0.8095238095238095,
854
+ "grad_norm": 0.22905386984348297,
855
+ "learning_rate": 0.0001717911923064442,
856
+ "loss": 0.1365,
857
+ "memory/device_mem_reserved(gib)": 47.13,
858
+ "memory/max_mem_active(gib)": 45.47,
859
+ "memory/max_mem_allocated(gib)": 45.47,
860
+ "step": 85
861
+ },
862
+ {
863
+ "epoch": 0.819047619047619,
864
+ "grad_norm": 0.26005449891090393,
865
+ "learning_rate": 0.00017107272551064473,
866
+ "loss": 0.1756,
867
+ "memory/device_mem_reserved(gib)": 47.13,
868
+ "memory/max_mem_active(gib)": 45.47,
869
+ "memory/max_mem_allocated(gib)": 45.47,
870
+ "step": 86
871
+ },
872
+ {
873
+ "epoch": 0.8285714285714286,
874
+ "grad_norm": 0.25042417645454407,
875
+ "learning_rate": 0.00017034676743453499,
876
+ "loss": 0.15,
877
+ "memory/device_mem_reserved(gib)": 47.13,
878
+ "memory/max_mem_active(gib)": 45.47,
879
+ "memory/max_mem_allocated(gib)": 45.47,
880
+ "step": 87
881
+ },
882
+ {
883
+ "epoch": 0.8380952380952381,
884
+ "grad_norm": 0.34358277916908264,
885
+ "learning_rate": 0.0001696133945962927,
886
+ "loss": 0.1763,
887
+ "memory/device_mem_reserved(gib)": 47.13,
888
+ "memory/max_mem_active(gib)": 45.47,
889
+ "memory/max_mem_allocated(gib)": 45.47,
890
+ "step": 88
891
+ },
892
+ {
893
+ "epoch": 0.8476190476190476,
894
+ "grad_norm": 0.23855257034301758,
895
+ "learning_rate": 0.0001688726842956339,
896
+ "loss": 0.1547,
897
+ "memory/device_mem_reserved(gib)": 47.13,
898
+ "memory/max_mem_active(gib)": 45.47,
899
+ "memory/max_mem_allocated(gib)": 45.47,
900
+ "step": 89
901
+ },
902
+ {
903
+ "epoch": 0.8571428571428571,
904
+ "grad_norm": 0.2579350173473358,
905
+ "learning_rate": 0.0001681247146056654,
906
+ "loss": 0.1872,
907
+ "memory/device_mem_reserved(gib)": 47.13,
908
+ "memory/max_mem_active(gib)": 45.47,
909
+ "memory/max_mem_allocated(gib)": 45.47,
910
+ "step": 90
911
+ },
912
+ {
913
+ "epoch": 0.8666666666666667,
914
+ "grad_norm": 0.30034390091896057,
915
+ "learning_rate": 0.00016736956436465573,
916
+ "loss": 0.1558,
917
+ "memory/device_mem_reserved(gib)": 47.13,
918
+ "memory/max_mem_active(gib)": 45.47,
919
+ "memory/max_mem_allocated(gib)": 45.47,
920
+ "step": 91
921
+ },
922
+ {
923
+ "epoch": 0.8761904761904762,
924
+ "grad_norm": 0.28185054659843445,
925
+ "learning_rate": 0.00016660731316772505,
926
+ "loss": 0.1878,
927
+ "memory/device_mem_reserved(gib)": 47.13,
928
+ "memory/max_mem_active(gib)": 45.47,
929
+ "memory/max_mem_allocated(gib)": 45.47,
930
+ "step": 92
931
+ },
932
+ {
933
+ "epoch": 0.8857142857142857,
934
+ "grad_norm": 0.32514333724975586,
935
+ "learning_rate": 0.0001658380413584558,
936
+ "loss": 0.1705,
937
+ "memory/device_mem_reserved(gib)": 47.13,
938
+ "memory/max_mem_active(gib)": 45.47,
939
+ "memory/max_mem_allocated(gib)": 45.47,
940
+ "step": 93
941
+ },
942
+ {
943
+ "epoch": 0.8952380952380953,
944
+ "grad_norm": 0.30903860926628113,
945
+ "learning_rate": 0.0001650618300204242,
946
+ "loss": 0.1708,
947
+ "memory/device_mem_reserved(gib)": 47.13,
948
+ "memory/max_mem_active(gib)": 45.47,
949
+ "memory/max_mem_allocated(gib)": 45.47,
950
+ "step": 94
951
+ },
952
+ {
953
+ "epoch": 0.9047619047619048,
954
+ "grad_norm": 0.24723806977272034,
955
+ "learning_rate": 0.00016427876096865394,
956
+ "loss": 0.1571,
957
+ "memory/device_mem_reserved(gib)": 47.13,
958
+ "memory/max_mem_active(gib)": 45.47,
959
+ "memory/max_mem_allocated(gib)": 45.47,
960
+ "step": 95
961
+ },
962
+ {
963
+ "epoch": 0.9142857142857143,
964
+ "grad_norm": 0.2727813422679901,
965
+ "learning_rate": 0.0001634889167409923,
966
+ "loss": 0.1779,
967
+ "memory/device_mem_reserved(gib)": 47.13,
968
+ "memory/max_mem_active(gib)": 45.47,
969
+ "memory/max_mem_allocated(gib)": 45.47,
970
+ "step": 96
971
+ },
972
+ {
973
+ "epoch": 0.9238095238095239,
974
+ "grad_norm": 0.2521083354949951,
975
+ "learning_rate": 0.0001626923805894107,
976
+ "loss": 0.1451,
977
+ "memory/device_mem_reserved(gib)": 47.13,
978
+ "memory/max_mem_active(gib)": 45.47,
979
+ "memory/max_mem_allocated(gib)": 45.47,
980
+ "step": 97
981
+ },
982
+ {
983
+ "epoch": 0.9333333333333333,
984
+ "grad_norm": 0.25517475605010986,
985
+ "learning_rate": 0.00016188923647122947,
986
+ "loss": 0.1789,
987
+ "memory/device_mem_reserved(gib)": 47.13,
988
+ "memory/max_mem_active(gib)": 45.47,
989
+ "memory/max_mem_allocated(gib)": 45.47,
990
+ "step": 98
991
+ },
992
+ {
993
+ "epoch": 0.9428571428571428,
994
+ "grad_norm": 0.26679396629333496,
995
+ "learning_rate": 0.0001610795690402688,
996
+ "loss": 0.1835,
997
+ "memory/device_mem_reserved(gib)": 47.13,
998
+ "memory/max_mem_active(gib)": 45.47,
999
+ "memory/max_mem_allocated(gib)": 45.47,
1000
+ "step": 99
1001
+ },
1002
+ {
1003
+ "epoch": 0.9523809523809523,
1004
+ "grad_norm": 0.22741852700710297,
1005
+ "learning_rate": 0.00016026346363792567,
1006
+ "loss": 0.1599,
1007
+ "memory/device_mem_reserved(gib)": 47.13,
1008
+ "memory/max_mem_active(gib)": 45.47,
1009
+ "memory/max_mem_allocated(gib)": 45.47,
1010
+ "step": 100
1011
+ },
1012
+ {
1013
+ "epoch": 0.9619047619047619,
1014
+ "grad_norm": 0.28691986203193665,
1015
+ "learning_rate": 0.00015944100628417868,
1016
+ "loss": 0.1654,
1017
+ "memory/device_mem_reserved(gib)": 47.13,
1018
+ "memory/max_mem_active(gib)": 45.47,
1019
+ "memory/max_mem_allocated(gib)": 45.47,
1020
+ "step": 101
1021
+ },
1022
+ {
1023
+ "epoch": 0.9714285714285714,
1024
+ "grad_norm": 0.24840302765369415,
1025
+ "learning_rate": 0.00015861228366852148,
1026
+ "loss": 0.1584,
1027
+ "memory/device_mem_reserved(gib)": 47.13,
1028
+ "memory/max_mem_active(gib)": 45.47,
1029
+ "memory/max_mem_allocated(gib)": 45.47,
1030
+ "step": 102
1031
+ },
1032
+ {
1033
+ "epoch": 0.9809523809523809,
1034
+ "grad_norm": 0.24804599583148956,
1035
+ "learning_rate": 0.00015777738314082514,
1036
+ "loss": 0.1391,
1037
+ "memory/device_mem_reserved(gib)": 47.13,
1038
+ "memory/max_mem_active(gib)": 45.47,
1039
+ "memory/max_mem_allocated(gib)": 45.47,
1040
+ "step": 103
1041
+ },
1042
+ {
1043
+ "epoch": 0.9904761904761905,
1044
+ "grad_norm": 0.2463735193014145,
1045
+ "learning_rate": 0.00015693639270213136,
1046
+ "loss": 0.1408,
1047
+ "memory/device_mem_reserved(gib)": 47.13,
1048
+ "memory/max_mem_active(gib)": 45.47,
1049
+ "memory/max_mem_allocated(gib)": 45.47,
1050
+ "step": 104
1051
+ },
1052
+ {
1053
+ "epoch": 1.0,
1054
+ "grad_norm": 0.29896217584609985,
1055
+ "learning_rate": 0.000156089400995377,
1056
+ "loss": 0.1666,
1057
+ "memory/device_mem_reserved(gib)": 47.13,
1058
+ "memory/max_mem_active(gib)": 45.47,
1059
+ "memory/max_mem_allocated(gib)": 45.47,
1060
+ "step": 105
1061
+ },
1062
+ {
1063
+ "epoch": 1.0095238095238095,
1064
+ "grad_norm": 0.2636161744594574,
1065
+ "learning_rate": 0.0001552364972960506,
1066
+ "loss": 0.1714,
1067
+ "memory/device_mem_reserved(gib)": 47.13,
1068
+ "memory/max_mem_active(gib)": 45.47,
1069
+ "memory/max_mem_allocated(gib)": 45.47,
1070
+ "step": 106
1071
+ },
1072
+ {
1073
+ "epoch": 1.019047619047619,
1074
+ "grad_norm": 0.2503117322921753,
1075
+ "learning_rate": 0.00015437777150278267,
1076
+ "loss": 0.1522,
1077
+ "memory/device_mem_reserved(gib)": 47.13,
1078
+ "memory/max_mem_active(gib)": 45.47,
1079
+ "memory/max_mem_allocated(gib)": 45.47,
1080
+ "step": 107
1081
+ },
1082
+ {
1083
+ "epoch": 1.0285714285714285,
1084
+ "grad_norm": 0.3032307028770447,
1085
+ "learning_rate": 0.00015351331412787004,
1086
+ "loss": 0.163,
1087
+ "memory/device_mem_reserved(gib)": 47.13,
1088
+ "memory/max_mem_active(gib)": 45.47,
1089
+ "memory/max_mem_allocated(gib)": 45.47,
1090
+ "step": 108
1091
+ },
1092
+ {
1093
+ "epoch": 1.0380952380952382,
1094
+ "grad_norm": 0.2652963399887085,
1095
+ "learning_rate": 0.0001526432162877356,
1096
+ "loss": 0.1589,
1097
+ "memory/device_mem_reserved(gib)": 47.13,
1098
+ "memory/max_mem_active(gib)": 45.47,
1099
+ "memory/max_mem_allocated(gib)": 45.47,
1100
+ "step": 109
1101
+ },
1102
+ {
1103
+ "epoch": 1.0476190476190477,
1104
+ "grad_norm": 0.2623184621334076,
1105
+ "learning_rate": 0.00015176756969332425,
1106
+ "loss": 0.1429,
1107
+ "memory/device_mem_reserved(gib)": 47.13,
1108
+ "memory/max_mem_active(gib)": 45.47,
1109
+ "memory/max_mem_allocated(gib)": 45.47,
1110
+ "step": 110
1111
+ },
1112
+ {
1113
+ "epoch": 1.0571428571428572,
1114
+ "grad_norm": 0.28222620487213135,
1115
+ "learning_rate": 0.0001508864666404365,
1116
+ "loss": 0.1469,
1117
+ "memory/device_mem_reserved(gib)": 47.13,
1118
+ "memory/max_mem_active(gib)": 45.47,
1119
+ "memory/max_mem_allocated(gib)": 45.47,
1120
+ "step": 111
1121
+ },
1122
+ {
1123
+ "epoch": 1.0666666666666667,
1124
+ "grad_norm": 0.28255829215049744,
1125
+ "learning_rate": 0.00015000000000000001,
1126
+ "loss": 0.156,
1127
+ "memory/device_mem_reserved(gib)": 47.13,
1128
+ "memory/max_mem_active(gib)": 45.47,
1129
+ "memory/max_mem_allocated(gib)": 45.47,
1130
+ "step": 112
1131
+ },
1132
+ {
1133
+ "epoch": 1.0761904761904761,
1134
+ "grad_norm": 0.22465772926807404,
1135
+ "learning_rate": 0.00014910826320828084,
1136
+ "loss": 0.131,
1137
+ "memory/device_mem_reserved(gib)": 47.13,
1138
+ "memory/max_mem_active(gib)": 45.47,
1139
+ "memory/max_mem_allocated(gib)": 45.47,
1140
+ "step": 113
1141
+ },
1142
+ {
1143
+ "epoch": 1.0857142857142856,
1144
+ "grad_norm": 0.30199363827705383,
1145
+ "learning_rate": 0.0001482113502570349,
1146
+ "loss": 0.1599,
1147
+ "memory/device_mem_reserved(gib)": 47.13,
1148
+ "memory/max_mem_active(gib)": 45.47,
1149
+ "memory/max_mem_allocated(gib)": 45.47,
1150
+ "step": 114
1151
+ },
1152
+ {
1153
+ "epoch": 1.0952380952380953,
1154
+ "grad_norm": 0.255288302898407,
1155
+ "learning_rate": 0.00014730935568360102,
1156
+ "loss": 0.1636,
1157
+ "memory/device_mem_reserved(gib)": 47.13,
1158
+ "memory/max_mem_active(gib)": 45.47,
1159
+ "memory/max_mem_allocated(gib)": 45.47,
1160
+ "step": 115
1161
+ },
1162
+ {
1163
+ "epoch": 1.1047619047619048,
1164
+ "grad_norm": 0.26142561435699463,
1165
+ "learning_rate": 0.00014640237456093634,
1166
+ "loss": 0.1518,
1167
+ "memory/device_mem_reserved(gib)": 47.13,
1168
+ "memory/max_mem_active(gib)": 45.47,
1169
+ "memory/max_mem_allocated(gib)": 45.47,
1170
+ "step": 116
1171
+ },
1172
+ {
1173
+ "epoch": 1.1142857142857143,
1174
+ "grad_norm": 0.30472439527511597,
1175
+ "learning_rate": 0.00014549050248759547,
1176
+ "loss": 0.1594,
1177
+ "memory/device_mem_reserved(gib)": 47.13,
1178
+ "memory/max_mem_active(gib)": 45.47,
1179
+ "memory/max_mem_allocated(gib)": 45.47,
1180
+ "step": 117
1181
+ },
1182
+ {
1183
+ "epoch": 1.1238095238095238,
1184
+ "grad_norm": 0.26955774426460266,
1185
+ "learning_rate": 0.00014457383557765386,
1186
+ "loss": 0.1676,
1187
+ "memory/device_mem_reserved(gib)": 47.13,
1188
+ "memory/max_mem_active(gib)": 45.47,
1189
+ "memory/max_mem_allocated(gib)": 45.47,
1190
+ "step": 118
1191
+ },
1192
+ {
1193
+ "epoch": 1.1333333333333333,
1194
+ "grad_norm": 0.34260597825050354,
1195
+ "learning_rate": 0.00014365247045057734,
1196
+ "loss": 0.1583,
1197
+ "memory/device_mem_reserved(gib)": 47.13,
1198
+ "memory/max_mem_active(gib)": 45.47,
1199
+ "memory/max_mem_allocated(gib)": 45.47,
1200
+ "step": 119
1201
+ },
1202
+ {
1203
+ "epoch": 1.1428571428571428,
1204
+ "grad_norm": 0.3136855661869049,
1205
+ "learning_rate": 0.0001427265042210381,
1206
+ "loss": 0.1558,
1207
+ "memory/device_mem_reserved(gib)": 47.13,
1208
+ "memory/max_mem_active(gib)": 45.47,
1209
+ "memory/max_mem_allocated(gib)": 45.47,
1210
+ "step": 120
1211
+ },
1212
+ {
1213
+ "epoch": 1.1523809523809523,
1214
+ "grad_norm": 0.27992814779281616,
1215
+ "learning_rate": 0.00014179603448867835,
1216
+ "loss": 0.1509,
1217
+ "memory/device_mem_reserved(gib)": 47.13,
1218
+ "memory/max_mem_active(gib)": 45.47,
1219
+ "memory/max_mem_allocated(gib)": 45.47,
1220
+ "step": 121
1221
+ },
1222
+ {
1223
+ "epoch": 1.161904761904762,
1224
+ "grad_norm": 0.34893789887428284,
1225
+ "learning_rate": 0.00014086115932782314,
1226
+ "loss": 0.1395,
1227
+ "memory/device_mem_reserved(gib)": 47.13,
1228
+ "memory/max_mem_active(gib)": 45.47,
1229
+ "memory/max_mem_allocated(gib)": 45.47,
1230
+ "step": 122
1231
+ },
1232
+ {
1233
+ "epoch": 1.1714285714285715,
1234
+ "grad_norm": 0.3308560252189636,
1235
+ "learning_rate": 0.0001399219772771431,
1236
+ "loss": 0.1591,
1237
+ "memory/device_mem_reserved(gib)": 47.13,
1238
+ "memory/max_mem_active(gib)": 45.47,
1239
+ "memory/max_mem_allocated(gib)": 45.47,
1240
+ "step": 123
1241
+ },
1242
+ {
1243
+ "epoch": 1.180952380952381,
1244
+ "grad_norm": 0.2736824154853821,
1245
+ "learning_rate": 0.00013897858732926793,
1246
+ "loss": 0.1604,
1247
+ "memory/device_mem_reserved(gib)": 47.13,
1248
+ "memory/max_mem_active(gib)": 45.47,
1249
+ "memory/max_mem_allocated(gib)": 45.47,
1250
+ "step": 124
1251
+ },
1252
+ {
1253
+ "epoch": 1.1904761904761905,
1254
+ "grad_norm": 0.28148430585861206,
1255
+ "learning_rate": 0.0001380310889203526,
1256
+ "loss": 0.1368,
1257
+ "memory/device_mem_reserved(gib)": 47.13,
1258
+ "memory/max_mem_active(gib)": 45.47,
1259
+ "memory/max_mem_allocated(gib)": 45.47,
1260
+ "step": 125
1261
+ },
1262
+ {
1263
+ "epoch": 1.2,
1264
+ "grad_norm": 0.25899434089660645,
1265
+ "learning_rate": 0.00013707958191959608,
1266
+ "loss": 0.15,
1267
+ "memory/device_mem_reserved(gib)": 47.13,
1268
+ "memory/max_mem_active(gib)": 45.47,
1269
+ "memory/max_mem_allocated(gib)": 45.47,
1270
+ "step": 126
1271
+ },
1272
+ {
1273
+ "epoch": 1.2095238095238094,
1274
+ "grad_norm": 0.32127872109413147,
1275
+ "learning_rate": 0.00013612416661871533,
1276
+ "loss": 0.1522,
1277
+ "memory/device_mem_reserved(gib)": 47.13,
1278
+ "memory/max_mem_active(gib)": 45.47,
1279
+ "memory/max_mem_allocated(gib)": 45.47,
1280
+ "step": 127
1281
+ },
1282
+ {
1283
+ "epoch": 1.2190476190476192,
1284
+ "grad_norm": 0.350427508354187,
1285
+ "learning_rate": 0.00013516494372137368,
1286
+ "loss": 0.1361,
1287
+ "memory/device_mem_reserved(gib)": 47.13,
1288
+ "memory/max_mem_active(gib)": 45.47,
1289
+ "memory/max_mem_allocated(gib)": 45.47,
1290
+ "step": 128
1291
+ },
1292
+ {
1293
+ "epoch": 1.2285714285714286,
1294
+ "grad_norm": 0.36603009700775146,
1295
+ "learning_rate": 0.00013420201433256689,
1296
+ "loss": 0.1527,
1297
+ "memory/device_mem_reserved(gib)": 47.13,
1298
+ "memory/max_mem_active(gib)": 45.47,
1299
+ "memory/max_mem_allocated(gib)": 45.47,
1300
+ "step": 129
1301
+ },
1302
+ {
1303
+ "epoch": 1.2380952380952381,
1304
+ "grad_norm": 0.2870017886161804,
1305
+ "learning_rate": 0.00013323547994796597,
1306
+ "loss": 0.152,
1307
+ "memory/device_mem_reserved(gib)": 47.13,
1308
+ "memory/max_mem_active(gib)": 45.47,
1309
+ "memory/max_mem_allocated(gib)": 45.47,
1310
+ "step": 130
1311
+ },
1312
+ {
1313
+ "epoch": 1.2476190476190476,
1314
+ "grad_norm": 0.29138877987861633,
1315
+ "learning_rate": 0.0001322654424432195,
1316
+ "loss": 0.1599,
1317
+ "memory/device_mem_reserved(gib)": 47.13,
1318
+ "memory/max_mem_active(gib)": 45.47,
1319
+ "memory/max_mem_allocated(gib)": 45.47,
1320
+ "step": 131
1321
+ },
1322
+ {
1323
+ "epoch": 1.2571428571428571,
1324
+ "grad_norm": 0.3358840048313141,
1325
+ "learning_rate": 0.00013129200406321545,
1326
+ "loss": 0.1668,
1327
+ "memory/device_mem_reserved(gib)": 47.13,
1328
+ "memory/max_mem_active(gib)": 45.47,
1329
+ "memory/max_mem_allocated(gib)": 45.47,
1330
+ "step": 132
1331
+ },
1332
+ {
1333
+ "epoch": 1.2666666666666666,
1334
+ "grad_norm": 0.3152446746826172,
1335
+ "learning_rate": 0.00013031526741130435,
1336
+ "loss": 0.1382,
1337
+ "memory/device_mem_reserved(gib)": 47.13,
1338
+ "memory/max_mem_active(gib)": 45.47,
1339
+ "memory/max_mem_allocated(gib)": 45.47,
1340
+ "step": 133
1341
+ },
1342
+ {
1343
+ "epoch": 1.276190476190476,
1344
+ "grad_norm": 0.2602708041667938,
1345
+ "learning_rate": 0.00012933533543848461,
1346
+ "loss": 0.1441,
1347
+ "memory/device_mem_reserved(gib)": 47.13,
1348
+ "memory/max_mem_active(gib)": 45.47,
1349
+ "memory/max_mem_allocated(gib)": 45.47,
1350
+ "step": 134
1351
+ },
1352
+ {
1353
+ "epoch": 1.2857142857142856,
1354
+ "grad_norm": 0.27083972096443176,
1355
+ "learning_rate": 0.0001283523114325511,
1356
+ "loss": 0.1517,
1357
+ "memory/device_mem_reserved(gib)": 47.13,
1358
+ "memory/max_mem_active(gib)": 45.47,
1359
+ "memory/max_mem_allocated(gib)": 45.47,
1360
+ "step": 135
1361
+ },
1362
+ {
1363
+ "epoch": 1.2952380952380953,
1364
+ "grad_norm": 0.27242034673690796,
1365
+ "learning_rate": 0.0001273662990072083,
1366
+ "loss": 0.1511,
1367
+ "memory/device_mem_reserved(gib)": 47.13,
1368
+ "memory/max_mem_active(gib)": 45.47,
1369
+ "memory/max_mem_allocated(gib)": 45.47,
1370
+ "step": 136
1371
+ },
1372
+ {
1373
+ "epoch": 1.3047619047619048,
1374
+ "grad_norm": 0.28473100066185,
1375
+ "learning_rate": 0.0001263774020911492,
1376
+ "loss": 0.1339,
1377
+ "memory/device_mem_reserved(gib)": 47.13,
1378
+ "memory/max_mem_active(gib)": 45.47,
1379
+ "memory/max_mem_allocated(gib)": 45.47,
1380
+ "step": 137
1381
+ },
1382
+ {
1383
+ "epoch": 1.3142857142857143,
1384
+ "grad_norm": 0.27939581871032715,
1385
+ "learning_rate": 0.0001253857249171008,
1386
+ "loss": 0.1414,
1387
+ "memory/device_mem_reserved(gib)": 47.13,
1388
+ "memory/max_mem_active(gib)": 45.47,
1389
+ "memory/max_mem_allocated(gib)": 45.47,
1390
+ "step": 138
1391
+ },
1392
+ {
1393
+ "epoch": 1.3238095238095238,
1394
+ "grad_norm": 0.2833685874938965,
1395
+ "learning_rate": 0.00012439137201083773,
1396
+ "loss": 0.139,
1397
+ "memory/device_mem_reserved(gib)": 47.13,
1398
+ "memory/max_mem_active(gib)": 45.47,
1399
+ "memory/max_mem_allocated(gib)": 45.47,
1400
+ "step": 139
1401
+ },
1402
+ {
1403
+ "epoch": 1.3333333333333333,
1404
+ "grad_norm": 0.3193693161010742,
1405
+ "learning_rate": 0.0001233944481801649,
1406
+ "loss": 0.159,
1407
+ "memory/device_mem_reserved(gib)": 47.13,
1408
+ "memory/max_mem_active(gib)": 45.47,
1409
+ "memory/max_mem_allocated(gib)": 45.47,
1410
+ "step": 140
1411
+ },
1412
+ {
1413
+ "epoch": 1.342857142857143,
1414
+ "grad_norm": 0.28556281328201294,
1415
+ "learning_rate": 0.0001223950585038703,
1416
+ "loss": 0.1549,
1417
+ "memory/device_mem_reserved(gib)": 47.13,
1418
+ "memory/max_mem_active(gib)": 45.47,
1419
+ "memory/max_mem_allocated(gib)": 45.47,
1420
+ "step": 141
1421
+ },
1422
+ {
1423
+ "epoch": 1.3523809523809525,
1424
+ "grad_norm": 0.2670241594314575,
1425
+ "learning_rate": 0.00012139330832064974,
1426
+ "loss": 0.1595,
1427
+ "memory/device_mem_reserved(gib)": 47.13,
1428
+ "memory/max_mem_active(gib)": 45.47,
1429
+ "memory/max_mem_allocated(gib)": 45.47,
1430
+ "step": 142
1431
+ },
1432
+ {
1433
+ "epoch": 1.361904761904762,
1434
+ "grad_norm": 0.20607773959636688,
1435
+ "learning_rate": 0.00012038930321800346,
1436
+ "loss": 0.1299,
1437
+ "memory/device_mem_reserved(gib)": 47.13,
1438
+ "memory/max_mem_active(gib)": 45.47,
1439
+ "memory/max_mem_allocated(gib)": 45.47,
1440
+ "step": 143
1441
+ },
1442
+ {
1443
+ "epoch": 1.3714285714285714,
1444
+ "grad_norm": 0.2973734736442566,
1445
+ "learning_rate": 0.00011938314902110701,
1446
+ "loss": 0.1552,
1447
+ "memory/device_mem_reserved(gib)": 47.13,
1448
+ "memory/max_mem_active(gib)": 45.47,
1449
+ "memory/max_mem_allocated(gib)": 45.47,
1450
+ "step": 144
1451
+ },
1452
+ {
1453
+ "epoch": 1.380952380952381,
1454
+ "grad_norm": 0.3032817244529724,
1455
+ "learning_rate": 0.00011837495178165706,
1456
+ "loss": 0.1554,
1457
+ "memory/device_mem_reserved(gib)": 47.13,
1458
+ "memory/max_mem_active(gib)": 45.47,
1459
+ "memory/max_mem_allocated(gib)": 45.47,
1460
+ "step": 145
1461
+ },
1462
+ {
1463
+ "epoch": 1.3904761904761904,
1464
+ "grad_norm": 0.2765556871891022,
1465
+ "learning_rate": 0.00011736481776669306,
1466
+ "loss": 0.1346,
1467
+ "memory/device_mem_reserved(gib)": 47.13,
1468
+ "memory/max_mem_active(gib)": 45.47,
1469
+ "memory/max_mem_allocated(gib)": 45.47,
1470
+ "step": 146
1471
+ },
1472
+ {
1473
+ "epoch": 1.4,
1474
+ "grad_norm": 0.24869301915168762,
1475
+ "learning_rate": 0.00011635285344739651,
1476
+ "loss": 0.1573,
1477
+ "memory/device_mem_reserved(gib)": 47.13,
1478
+ "memory/max_mem_active(gib)": 45.47,
1479
+ "memory/max_mem_allocated(gib)": 45.47,
1480
+ "step": 147
1481
+ },
1482
+ {
1483
+ "epoch": 1.4095238095238094,
1484
+ "grad_norm": 0.28329774737358093,
1485
+ "learning_rate": 0.00011533916548786857,
1486
+ "loss": 0.166,
1487
+ "memory/device_mem_reserved(gib)": 47.13,
1488
+ "memory/max_mem_active(gib)": 45.47,
1489
+ "memory/max_mem_allocated(gib)": 45.47,
1490
+ "step": 148
1491
+ },
1492
+ {
1493
+ "epoch": 1.4190476190476191,
1494
+ "grad_norm": 0.31719744205474854,
1495
+ "learning_rate": 0.00011432386073388717,
1496
+ "loss": 0.1556,
1497
+ "memory/device_mem_reserved(gib)": 47.13,
1498
+ "memory/max_mem_active(gib)": 45.47,
1499
+ "memory/max_mem_allocated(gib)": 45.47,
1500
+ "step": 149
1501
+ },
1502
+ {
1503
+ "epoch": 1.4285714285714286,
1504
+ "grad_norm": 0.27932024002075195,
1505
+ "learning_rate": 0.00011330704620164538,
1506
+ "loss": 0.1722,
1507
+ "memory/device_mem_reserved(gib)": 47.13,
1508
+ "memory/max_mem_active(gib)": 45.47,
1509
+ "memory/max_mem_allocated(gib)": 45.47,
1510
+ "step": 150
1511
+ },
1512
+ {
1513
+ "epoch": 1.438095238095238,
1514
+ "grad_norm": 0.2447134554386139,
1515
+ "learning_rate": 0.00011228882906647142,
1516
+ "loss": 0.1318,
1517
+ "memory/device_mem_reserved(gib)": 47.13,
1518
+ "memory/max_mem_active(gib)": 45.47,
1519
+ "memory/max_mem_allocated(gib)": 45.47,
1520
+ "step": 151
1521
+ },
1522
+ {
1523
+ "epoch": 1.4476190476190476,
1524
+ "grad_norm": 0.2684175670146942,
1525
+ "learning_rate": 0.00011126931665153212,
1526
+ "loss": 0.1465,
1527
+ "memory/device_mem_reserved(gib)": 47.13,
1528
+ "memory/max_mem_active(gib)": 45.47,
1529
+ "memory/max_mem_allocated(gib)": 45.47,
1530
+ "step": 152
1531
+ },
1532
+ {
1533
+ "epoch": 1.457142857142857,
1534
+ "grad_norm": 0.24706321954727173,
1535
+ "learning_rate": 0.00011024861641652071,
1536
+ "loss": 0.1226,
1537
+ "memory/device_mem_reserved(gib)": 47.13,
1538
+ "memory/max_mem_active(gib)": 45.47,
1539
+ "memory/max_mem_allocated(gib)": 45.47,
1540
+ "step": 153
1541
+ },
1542
+ {
1543
+ "epoch": 1.4666666666666668,
1544
+ "grad_norm": 0.2958153486251831,
1545
+ "learning_rate": 0.00010922683594633021,
1546
+ "loss": 0.1447,
1547
+ "memory/device_mem_reserved(gib)": 47.13,
1548
+ "memory/max_mem_active(gib)": 45.47,
1549
+ "memory/max_mem_allocated(gib)": 45.47,
1550
+ "step": 154
1551
+ },
1552
+ {
1553
+ "epoch": 1.4761904761904763,
1554
+ "grad_norm": 0.305876761674881,
1555
+ "learning_rate": 0.00010820408293971378,
1556
+ "loss": 0.1566,
1557
+ "memory/device_mem_reserved(gib)": 47.13,
1558
+ "memory/max_mem_active(gib)": 45.47,
1559
+ "memory/max_mem_allocated(gib)": 45.47,
1560
+ "step": 155
1561
+ },
1562
+ {
1563
+ "epoch": 1.4857142857142858,
1564
+ "grad_norm": 0.2755098342895508,
1565
+ "learning_rate": 0.00010718046519793276,
1566
+ "loss": 0.155,
1567
+ "memory/device_mem_reserved(gib)": 47.13,
1568
+ "memory/max_mem_active(gib)": 45.47,
1569
+ "memory/max_mem_allocated(gib)": 45.47,
1570
+ "step": 156
1571
+ },
1572
+ {
1573
+ "epoch": 1.4952380952380953,
1574
+ "grad_norm": 0.2550102174282074,
1575
+ "learning_rate": 0.00010615609061339432,
1576
+ "loss": 0.1227,
1577
+ "memory/device_mem_reserved(gib)": 47.13,
1578
+ "memory/max_mem_active(gib)": 45.47,
1579
+ "memory/max_mem_allocated(gib)": 45.47,
1580
+ "step": 157
1581
+ },
1582
+ {
1583
+ "epoch": 1.5047619047619047,
1584
+ "grad_norm": 0.26932892203330994,
1585
+ "learning_rate": 0.00010513106715827896,
1586
+ "loss": 0.1533,
1587
+ "memory/device_mem_reserved(gib)": 47.13,
1588
+ "memory/max_mem_active(gib)": 45.47,
1589
+ "memory/max_mem_allocated(gib)": 45.47,
1590
+ "step": 158
1591
+ },
1592
+ {
1593
+ "epoch": 1.5142857142857142,
1594
+ "grad_norm": 0.24724788963794708,
1595
+ "learning_rate": 0.00010410550287316034,
1596
+ "loss": 0.1346,
1597
+ "memory/device_mem_reserved(gib)": 47.13,
1598
+ "memory/max_mem_active(gib)": 45.47,
1599
+ "memory/max_mem_allocated(gib)": 45.47,
1600
+ "step": 159
1601
+ },
1602
+ {
1603
+ "epoch": 1.5238095238095237,
1604
+ "grad_norm": 0.2605031132698059,
1605
+ "learning_rate": 0.00010307950585561706,
1606
+ "loss": 0.1319,
1607
+ "memory/device_mem_reserved(gib)": 47.13,
1608
+ "memory/max_mem_active(gib)": 45.47,
1609
+ "memory/max_mem_allocated(gib)": 45.47,
1610
+ "step": 160
1611
+ },
1612
+ {
1613
+ "epoch": 1.5333333333333332,
1614
+ "grad_norm": 0.28353723883628845,
1615
+ "learning_rate": 0.00010205318424883905,
1616
+ "loss": 0.1386,
1617
+ "memory/device_mem_reserved(gib)": 47.13,
1618
+ "memory/max_mem_active(gib)": 45.47,
1619
+ "memory/max_mem_allocated(gib)": 45.47,
1620
+ "step": 161
1621
+ },
1622
+ {
1623
+ "epoch": 1.5428571428571427,
1624
+ "grad_norm": 0.31240761280059814,
1625
+ "learning_rate": 0.00010102664623022899,
1626
+ "loss": 0.1468,
1627
+ "memory/device_mem_reserved(gib)": 47.13,
1628
+ "memory/max_mem_active(gib)": 45.47,
1629
+ "memory/max_mem_allocated(gib)": 45.47,
1630
+ "step": 162
1631
+ },
1632
+ {
1633
+ "epoch": 1.5523809523809524,
1634
+ "grad_norm": 0.2614494562149048,
1635
+ "learning_rate": 0.0001,
1636
+ "loss": 0.146,
1637
+ "memory/device_mem_reserved(gib)": 47.13,
1638
+ "memory/max_mem_active(gib)": 45.47,
1639
+ "memory/max_mem_allocated(gib)": 45.47,
1640
+ "step": 163
1641
+ },
1642
+ {
1643
+ "epoch": 1.561904761904762,
1644
+ "grad_norm": 0.28343483805656433,
1645
+ "learning_rate": 9.897335376977102e-05,
1646
+ "loss": 0.1531,
1647
+ "memory/device_mem_reserved(gib)": 47.13,
1648
+ "memory/max_mem_active(gib)": 45.48,
1649
+ "memory/max_mem_allocated(gib)": 45.48,
1650
+ "step": 164
1651
+ },
1652
+ {
1653
+ "epoch": 1.5714285714285714,
1654
+ "grad_norm": 0.2741709053516388,
1655
+ "learning_rate": 9.794681575116097e-05,
1656
+ "loss": 0.1304,
1657
+ "memory/device_mem_reserved(gib)": 47.13,
1658
+ "memory/max_mem_active(gib)": 45.48,
1659
+ "memory/max_mem_allocated(gib)": 45.48,
1660
+ "step": 165
1661
+ },
1662
+ {
1663
+ "epoch": 1.580952380952381,
1664
+ "grad_norm": 0.326215922832489,
1665
+ "learning_rate": 9.692049414438299e-05,
1666
+ "loss": 0.1551,
1667
+ "memory/device_mem_reserved(gib)": 47.13,
1668
+ "memory/max_mem_active(gib)": 45.48,
1669
+ "memory/max_mem_allocated(gib)": 45.48,
1670
+ "step": 166
1671
+ },
1672
+ {
1673
+ "epoch": 1.5904761904761906,
1674
+ "grad_norm": 0.3120001256465912,
1675
+ "learning_rate": 9.58944971268397e-05,
1676
+ "loss": 0.1387,
1677
+ "memory/device_mem_reserved(gib)": 47.13,
1678
+ "memory/max_mem_active(gib)": 45.48,
1679
+ "memory/max_mem_allocated(gib)": 45.48,
1680
+ "step": 167
1681
+ },
1682
+ {
1683
+ "epoch": 1.6,
1684
+ "grad_norm": 0.2673357129096985,
1685
+ "learning_rate": 9.486893284172102e-05,
1686
+ "loss": 0.1389,
1687
+ "memory/device_mem_reserved(gib)": 47.13,
1688
+ "memory/max_mem_active(gib)": 45.48,
1689
+ "memory/max_mem_allocated(gib)": 45.48,
1690
+ "step": 168
1691
+ },
1692
+ {
1693
+ "epoch": 1.6095238095238096,
1694
+ "grad_norm": 0.3184024691581726,
1695
+ "learning_rate": 9.384390938660572e-05,
1696
+ "loss": 0.1388,
1697
+ "memory/device_mem_reserved(gib)": 47.13,
1698
+ "memory/max_mem_active(gib)": 45.48,
1699
+ "memory/max_mem_allocated(gib)": 45.48,
1700
+ "step": 169
1701
+ },
1702
+ {
1703
+ "epoch": 1.619047619047619,
1704
+ "grad_norm": 0.28744614124298096,
1705
+ "learning_rate": 9.281953480206725e-05,
1706
+ "loss": 0.1246,
1707
+ "memory/device_mem_reserved(gib)": 47.13,
1708
+ "memory/max_mem_active(gib)": 45.48,
1709
+ "memory/max_mem_allocated(gib)": 45.48,
1710
+ "step": 170
1711
+ },
1712
+ {
1713
+ "epoch": 1.6285714285714286,
1714
+ "grad_norm": 0.28229352831840515,
1715
+ "learning_rate": 9.179591706028626e-05,
1716
+ "loss": 0.1637,
1717
+ "memory/device_mem_reserved(gib)": 47.13,
1718
+ "memory/max_mem_active(gib)": 45.48,
1719
+ "memory/max_mem_allocated(gib)": 45.48,
1720
+ "step": 171
1721
+ },
1722
+ {
1723
+ "epoch": 1.638095238095238,
1724
+ "grad_norm": 0.2750794589519501,
1725
+ "learning_rate": 9.077316405366981e-05,
1726
+ "loss": 0.1176,
1727
+ "memory/device_mem_reserved(gib)": 47.13,
1728
+ "memory/max_mem_active(gib)": 45.48,
1729
+ "memory/max_mem_allocated(gib)": 45.48,
1730
+ "step": 172
1731
+ },
1732
+ {
1733
+ "epoch": 1.6476190476190475,
1734
+ "grad_norm": 0.28753307461738586,
1735
+ "learning_rate": 8.975138358347931e-05,
1736
+ "loss": 0.1517,
1737
+ "memory/device_mem_reserved(gib)": 47.13,
1738
+ "memory/max_mem_active(gib)": 45.48,
1739
+ "memory/max_mem_allocated(gib)": 45.48,
1740
+ "step": 173
1741
+ },
1742
+ {
1743
+ "epoch": 1.657142857142857,
1744
+ "grad_norm": 0.2737794518470764,
1745
+ "learning_rate": 8.87306833484679e-05,
1746
+ "loss": 0.1604,
1747
+ "memory/device_mem_reserved(gib)": 47.13,
1748
+ "memory/max_mem_active(gib)": 45.48,
1749
+ "memory/max_mem_allocated(gib)": 45.48,
1750
+ "step": 174
1751
+ },
1752
+ {
1753
+ "epoch": 1.6666666666666665,
1754
+ "grad_norm": 0.2456161379814148,
1755
+ "learning_rate": 8.77111709335286e-05,
1756
+ "loss": 0.1441,
1757
+ "memory/device_mem_reserved(gib)": 47.13,
1758
+ "memory/max_mem_active(gib)": 45.48,
1759
+ "memory/max_mem_allocated(gib)": 45.48,
1760
+ "step": 175
1761
+ },
1762
+ {
1763
+ "epoch": 1.6761904761904762,
1764
+ "grad_norm": 0.2895718812942505,
1765
+ "learning_rate": 8.669295379835467e-05,
1766
+ "loss": 0.152,
1767
+ "memory/device_mem_reserved(gib)": 47.13,
1768
+ "memory/max_mem_active(gib)": 45.48,
1769
+ "memory/max_mem_allocated(gib)": 45.48,
1770
+ "step": 176
1771
+ },
1772
+ {
1773
+ "epoch": 1.6857142857142857,
1774
+ "grad_norm": 0.3128141462802887,
1775
+ "learning_rate": 8.567613926611288e-05,
1776
+ "loss": 0.1692,
1777
+ "memory/device_mem_reserved(gib)": 47.13,
1778
+ "memory/max_mem_active(gib)": 45.48,
1779
+ "memory/max_mem_allocated(gib)": 45.48,
1780
+ "step": 177
1781
+ },
1782
+ {
1783
+ "epoch": 1.6952380952380952,
1784
+ "grad_norm": 0.29811859130859375,
1785
+ "learning_rate": 8.466083451213144e-05,
1786
+ "loss": 0.1351,
1787
+ "memory/device_mem_reserved(gib)": 47.13,
1788
+ "memory/max_mem_active(gib)": 45.48,
1789
+ "memory/max_mem_allocated(gib)": 45.48,
1790
+ "step": 178
1791
+ },
1792
+ {
1793
+ "epoch": 1.704761904761905,
1794
+ "grad_norm": 0.2686918377876282,
1795
+ "learning_rate": 8.364714655260349e-05,
1796
+ "loss": 0.1318,
1797
+ "memory/device_mem_reserved(gib)": 47.13,
1798
+ "memory/max_mem_active(gib)": 45.48,
1799
+ "memory/max_mem_allocated(gib)": 45.48,
1800
+ "step": 179
1801
+ },
1802
+ {
1803
+ "epoch": 1.7142857142857144,
1804
+ "grad_norm": 0.31242185831069946,
1805
+ "learning_rate": 8.263518223330697e-05,
1806
+ "loss": 0.1512,
1807
+ "memory/device_mem_reserved(gib)": 47.13,
1808
+ "memory/max_mem_active(gib)": 45.48,
1809
+ "memory/max_mem_allocated(gib)": 45.48,
1810
+ "step": 180
1811
+ },
1812
+ {
1813
+ "epoch": 1.723809523809524,
1814
+ "grad_norm": 0.3040373623371124,
1815
+ "learning_rate": 8.162504821834295e-05,
1816
+ "loss": 0.1464,
1817
+ "memory/device_mem_reserved(gib)": 47.13,
1818
+ "memory/max_mem_active(gib)": 45.48,
1819
+ "memory/max_mem_allocated(gib)": 45.48,
1820
+ "step": 181
1821
+ },
1822
+ {
1823
+ "epoch": 1.7333333333333334,
1824
+ "grad_norm": 0.2847082018852234,
1825
+ "learning_rate": 8.0616850978893e-05,
1826
+ "loss": 0.1586,
1827
+ "memory/device_mem_reserved(gib)": 47.13,
1828
+ "memory/max_mem_active(gib)": 45.48,
1829
+ "memory/max_mem_allocated(gib)": 45.48,
1830
+ "step": 182
1831
+ },
1832
+ {
1833
+ "epoch": 1.7428571428571429,
1834
+ "grad_norm": 0.27451273798942566,
1835
+ "learning_rate": 7.961069678199658e-05,
1836
+ "loss": 0.1553,
1837
+ "memory/device_mem_reserved(gib)": 47.13,
1838
+ "memory/max_mem_active(gib)": 45.48,
1839
+ "memory/max_mem_allocated(gib)": 45.48,
1840
+ "step": 183
1841
+ },
1842
+ {
1843
+ "epoch": 1.7523809523809524,
1844
+ "grad_norm": 0.3282047212123871,
1845
+ "learning_rate": 7.860669167935028e-05,
1846
+ "loss": 0.1559,
1847
+ "memory/device_mem_reserved(gib)": 47.13,
1848
+ "memory/max_mem_active(gib)": 45.48,
1849
+ "memory/max_mem_allocated(gib)": 45.48,
1850
+ "step": 184
1851
+ },
1852
+ {
1853
+ "epoch": 1.7619047619047619,
1854
+ "grad_norm": 0.2895129919052124,
1855
+ "learning_rate": 7.760494149612971e-05,
1856
+ "loss": 0.1431,
1857
+ "memory/device_mem_reserved(gib)": 47.13,
1858
+ "memory/max_mem_active(gib)": 45.48,
1859
+ "memory/max_mem_allocated(gib)": 45.48,
1860
+ "step": 185
1861
+ },
1862
+ {
1863
+ "epoch": 1.7714285714285714,
1864
+ "grad_norm": 0.2585572898387909,
1865
+ "learning_rate": 7.660555181983518e-05,
1866
+ "loss": 0.1274,
1867
+ "memory/device_mem_reserved(gib)": 47.13,
1868
+ "memory/max_mem_active(gib)": 45.48,
1869
+ "memory/max_mem_allocated(gib)": 45.48,
1870
+ "step": 186
1871
+ },
1872
+ {
1873
+ "epoch": 1.7809523809523808,
1874
+ "grad_norm": 0.2550938129425049,
1875
+ "learning_rate": 7.560862798916228e-05,
1876
+ "loss": 0.1482,
1877
+ "memory/device_mem_reserved(gib)": 47.13,
1878
+ "memory/max_mem_active(gib)": 45.48,
1879
+ "memory/max_mem_allocated(gib)": 45.48,
1880
+ "step": 187
1881
+ },
1882
+ {
1883
+ "epoch": 1.7904761904761903,
1884
+ "grad_norm": 0.267459511756897,
1885
+ "learning_rate": 7.461427508289922e-05,
1886
+ "loss": 0.1425,
1887
+ "memory/device_mem_reserved(gib)": 47.13,
1888
+ "memory/max_mem_active(gib)": 45.48,
1889
+ "memory/max_mem_allocated(gib)": 45.48,
1890
+ "step": 188
1891
+ },
1892
+ {
1893
+ "epoch": 1.8,
1894
+ "grad_norm": 0.273129940032959,
1895
+ "learning_rate": 7.36225979088508e-05,
1896
+ "loss": 0.118,
1897
+ "memory/device_mem_reserved(gib)": 47.13,
1898
+ "memory/max_mem_active(gib)": 45.48,
1899
+ "memory/max_mem_allocated(gib)": 45.48,
1900
+ "step": 189
1901
+ },
1902
+ {
1903
+ "epoch": 1.8095238095238095,
1904
+ "grad_norm": 0.33791351318359375,
1905
+ "learning_rate": 7.263370099279172e-05,
1906
+ "loss": 0.146,
1907
+ "memory/device_mem_reserved(gib)": 47.13,
1908
+ "memory/max_mem_active(gib)": 45.48,
1909
+ "memory/max_mem_allocated(gib)": 45.48,
1910
+ "step": 190
1911
+ },
1912
+ {
1913
+ "epoch": 1.819047619047619,
1914
+ "grad_norm": 0.3014010488986969,
1915
+ "learning_rate": 7.164768856744892e-05,
1916
+ "loss": 0.1461,
1917
+ "memory/device_mem_reserved(gib)": 47.13,
1918
+ "memory/max_mem_active(gib)": 45.48,
1919
+ "memory/max_mem_allocated(gib)": 45.48,
1920
+ "step": 191
1921
+ },
1922
+ {
1923
+ "epoch": 1.8285714285714287,
1924
+ "grad_norm": 0.32012397050857544,
1925
+ "learning_rate": 7.066466456151541e-05,
1926
+ "loss": 0.1782,
1927
+ "memory/device_mem_reserved(gib)": 47.13,
1928
+ "memory/max_mem_active(gib)": 45.48,
1929
+ "memory/max_mem_allocated(gib)": 45.48,
1930
+ "step": 192
1931
+ },
1932
+ {
1933
+ "epoch": 1.8380952380952382,
1934
+ "grad_norm": 0.25507038831710815,
1935
+ "learning_rate": 6.968473258869566e-05,
1936
+ "loss": 0.1319,
1937
+ "memory/device_mem_reserved(gib)": 47.13,
1938
+ "memory/max_mem_active(gib)": 45.48,
1939
+ "memory/max_mem_allocated(gib)": 45.48,
1940
+ "step": 193
1941
+ },
1942
+ {
1943
+ "epoch": 1.8476190476190477,
1944
+ "grad_norm": 0.28470760583877563,
1945
+ "learning_rate": 6.870799593678459e-05,
1946
+ "loss": 0.159,
1947
+ "memory/device_mem_reserved(gib)": 47.13,
1948
+ "memory/max_mem_active(gib)": 45.48,
1949
+ "memory/max_mem_allocated(gib)": 45.48,
1950
+ "step": 194
1951
+ },
1952
+ {
1953
+ "epoch": 1.8571428571428572,
1954
+ "grad_norm": 0.25228601694107056,
1955
+ "learning_rate": 6.773455755678054e-05,
1956
+ "loss": 0.1284,
1957
+ "memory/device_mem_reserved(gib)": 47.13,
1958
+ "memory/max_mem_active(gib)": 45.48,
1959
+ "memory/max_mem_allocated(gib)": 45.48,
1960
+ "step": 195
1961
+ },
1962
+ {
1963
+ "epoch": 1.8666666666666667,
1964
+ "grad_norm": 0.28710344433784485,
1965
+ "learning_rate": 6.676452005203406e-05,
1966
+ "loss": 0.1338,
1967
+ "memory/device_mem_reserved(gib)": 47.13,
1968
+ "memory/max_mem_active(gib)": 45.48,
1969
+ "memory/max_mem_allocated(gib)": 45.48,
1970
+ "step": 196
1971
+ },
1972
+ {
1973
+ "epoch": 1.8761904761904762,
1974
+ "grad_norm": 0.29423221945762634,
1975
+ "learning_rate": 6.579798566743314e-05,
1976
+ "loss": 0.1335,
1977
+ "memory/device_mem_reserved(gib)": 47.13,
1978
+ "memory/max_mem_active(gib)": 45.48,
1979
+ "memory/max_mem_allocated(gib)": 45.48,
1980
+ "step": 197
1981
+ },
1982
+ {
1983
+ "epoch": 1.8857142857142857,
1984
+ "grad_norm": 0.31776273250579834,
1985
+ "learning_rate": 6.483505627862632e-05,
1986
+ "loss": 0.1509,
1987
+ "memory/device_mem_reserved(gib)": 47.13,
1988
+ "memory/max_mem_active(gib)": 45.48,
1989
+ "memory/max_mem_allocated(gib)": 45.48,
1990
+ "step": 198
1991
+ },
1992
+ {
1993
+ "epoch": 1.8952380952380952,
1994
+ "grad_norm": 0.2739746868610382,
1995
+ "learning_rate": 6.387583338128471e-05,
1996
+ "loss": 0.1195,
1997
+ "memory/device_mem_reserved(gib)": 47.13,
1998
+ "memory/max_mem_active(gib)": 45.48,
1999
+ "memory/max_mem_allocated(gib)": 45.48,
2000
+ "step": 199
2001
+ },
2002
+ {
2003
+ "epoch": 1.9047619047619047,
2004
+ "grad_norm": 0.28000062704086304,
2005
+ "learning_rate": 6.292041808040393e-05,
2006
+ "loss": 0.1444,
2007
+ "memory/device_mem_reserved(gib)": 47.13,
2008
+ "memory/max_mem_active(gib)": 45.48,
2009
+ "memory/max_mem_allocated(gib)": 45.48,
2010
+ "step": 200
2011
+ },
2012
+ {
2013
+ "epoch": 1.9142857142857141,
2014
+ "grad_norm": 0.2637692987918854,
2015
+ "learning_rate": 6.196891107964744e-05,
2016
+ "loss": 0.1412,
2017
+ "memory/device_mem_reserved(gib)": 47.13,
2018
+ "memory/max_mem_active(gib)": 45.48,
2019
+ "memory/max_mem_allocated(gib)": 45.48,
2020
+ "step": 201
2021
+ },
2022
+ {
2023
+ "epoch": 1.9238095238095239,
2024
+ "grad_norm": 0.2977365255355835,
2025
+ "learning_rate": 6.102141267073207e-05,
2026
+ "loss": 0.1224,
2027
+ "memory/device_mem_reserved(gib)": 47.13,
2028
+ "memory/max_mem_active(gib)": 45.48,
2029
+ "memory/max_mem_allocated(gib)": 45.48,
2030
+ "step": 202
2031
+ },
2032
+ {
2033
+ "epoch": 1.9333333333333333,
2034
+ "grad_norm": 0.2656545042991638,
2035
+ "learning_rate": 6.007802272285693e-05,
2036
+ "loss": 0.1235,
2037
+ "memory/device_mem_reserved(gib)": 47.13,
2038
+ "memory/max_mem_active(gib)": 45.48,
2039
+ "memory/max_mem_allocated(gib)": 45.48,
2040
+ "step": 203
2041
+ },
2042
+ {
2043
+ "epoch": 1.9428571428571428,
2044
+ "grad_norm": 0.3028334975242615,
2045
+ "learning_rate": 5.913884067217685e-05,
2046
+ "loss": 0.1544,
2047
+ "memory/device_mem_reserved(gib)": 47.13,
2048
+ "memory/max_mem_active(gib)": 45.48,
2049
+ "memory/max_mem_allocated(gib)": 45.48,
2050
+ "step": 204
2051
+ },
2052
+ {
2053
+ "epoch": 1.9523809523809523,
2054
+ "grad_norm": 0.32587432861328125,
2055
+ "learning_rate": 5.82039655113217e-05,
2056
+ "loss": 0.1484,
2057
+ "memory/device_mem_reserved(gib)": 47.13,
2058
+ "memory/max_mem_active(gib)": 45.48,
2059
+ "memory/max_mem_allocated(gib)": 45.48,
2060
+ "step": 205
2061
+ },
2062
+ {
2063
+ "epoch": 1.961904761904762,
2064
+ "grad_norm": 0.2999066710472107,
2065
+ "learning_rate": 5.727349577896194e-05,
2066
+ "loss": 0.1392,
2067
+ "memory/device_mem_reserved(gib)": 47.13,
2068
+ "memory/max_mem_active(gib)": 45.48,
2069
+ "memory/max_mem_allocated(gib)": 45.48,
2070
+ "step": 206
2071
+ },
2072
+ {
2073
+ "epoch": 1.9714285714285715,
2074
+ "grad_norm": 0.2822760045528412,
2075
+ "learning_rate": 5.634752954942264e-05,
2076
+ "loss": 0.1334,
2077
+ "memory/device_mem_reserved(gib)": 47.13,
2078
+ "memory/max_mem_active(gib)": 45.48,
2079
+ "memory/max_mem_allocated(gib)": 45.48,
2080
+ "step": 207
2081
+ },
2082
+ {
2083
+ "epoch": 1.980952380952381,
2084
+ "grad_norm": 0.4100743234157562,
2085
+ "learning_rate": 5.542616442234618e-05,
2086
+ "loss": 0.1295,
2087
+ "memory/device_mem_reserved(gib)": 47.13,
2088
+ "memory/max_mem_active(gib)": 45.48,
2089
+ "memory/max_mem_allocated(gib)": 45.48,
2090
+ "step": 208
2091
+ },
2092
+ {
2093
+ "epoch": 1.9904761904761905,
2094
+ "grad_norm": 0.22527122497558594,
2095
+ "learning_rate": 5.450949751240456e-05,
2096
+ "loss": 0.1377,
2097
+ "memory/device_mem_reserved(gib)": 47.13,
2098
+ "memory/max_mem_active(gib)": 45.48,
2099
+ "memory/max_mem_allocated(gib)": 45.48,
2100
+ "step": 209
2101
+ },
2102
+ {
2103
+ "epoch": 2.0,
2104
+ "grad_norm": 0.33393946290016174,
2105
+ "learning_rate": 5.359762543906368e-05,
2106
+ "loss": 0.1251,
2107
+ "memory/device_mem_reserved(gib)": 47.13,
2108
+ "memory/max_mem_active(gib)": 45.48,
2109
+ "memory/max_mem_allocated(gib)": 45.48,
2110
+ "step": 210
2111
+ }
2112
+ ],
2113
+ "logging_steps": 1,
2114
+ "max_steps": 315,
2115
+ "num_input_tokens_seen": 0,
2116
+ "num_train_epochs": 3,
2117
+ "save_steps": 500,
2118
+ "stateful_callbacks": {
2119
+ "TrainerControl": {
2120
+ "args": {
2121
+ "should_epoch_stop": false,
2122
+ "should_evaluate": false,
2123
+ "should_log": false,
2124
+ "should_save": true,
2125
+ "should_training_stop": false
2126
+ },
2127
+ "attributes": {}
2128
+ }
2129
+ },
2130
+ "total_flos": 3.236037886423859e+16,
2131
+ "train_batch_size": 16,
2132
+ "trial_name": null,
2133
+ "trial_params": null
2134
+ }
checkpoint-210/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:232e77a355a770b00d79212785888b9356c3ed6653e1b633e2e2153e01eae951
3
+ size 6776
checkpoint-210/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-315/README.md ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B-Instruct-2507
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - axolotl
7
+ - base_model:adapter:Qwen/Qwen3-4B-Instruct-2507
8
+ - lora
9
+ - transformers
10
+ ---
11
+
12
+ # Model Card for Model ID
13
+
14
+ <!-- Provide a quick summary of what the model is/does. -->
15
+
16
+
17
+
18
+ ## Model Details
19
+
20
+ ### Model Description
21
+
22
+ <!-- Provide a longer summary of what this model is. -->
23
+
24
+
25
+
26
+ - **Developed by:** [More Information Needed]
27
+ - **Funded by [optional]:** [More Information Needed]
28
+ - **Shared by [optional]:** [More Information Needed]
29
+ - **Model type:** [More Information Needed]
30
+ - **Language(s) (NLP):** [More Information Needed]
31
+ - **License:** [More Information Needed]
32
+ - **Finetuned from model [optional]:** [More Information Needed]
33
+
34
+ ### Model Sources [optional]
35
+
36
+ <!-- Provide the basic links for the model. -->
37
+
38
+ - **Repository:** [More Information Needed]
39
+ - **Paper [optional]:** [More Information Needed]
40
+ - **Demo [optional]:** [More Information Needed]
41
+
42
+ ## Uses
43
+
44
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
45
+
46
+ ### Direct Use
47
+
48
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Downstream Use [optional]
53
+
54
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
55
+
56
+ [More Information Needed]
57
+
58
+ ### Out-of-Scope Use
59
+
60
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ## Bias, Risks, and Limitations
65
+
66
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
67
+
68
+ [More Information Needed]
69
+
70
+ ### Recommendations
71
+
72
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
73
+
74
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
75
+
76
+ ## How to Get Started with the Model
77
+
78
+ Use the code below to get started with the model.
79
+
80
+ [More Information Needed]
81
+
82
+ ## Training Details
83
+
84
+ ### Training Data
85
+
86
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
87
+
88
+ [More Information Needed]
89
+
90
+ ### Training Procedure
91
+
92
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
93
+
94
+ #### Preprocessing [optional]
95
+
96
+ [More Information Needed]
97
+
98
+
99
+ #### Training Hyperparameters
100
+
101
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
102
+
103
+ #### Speeds, Sizes, Times [optional]
104
+
105
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
106
+
107
+ [More Information Needed]
108
+
109
+ ## Evaluation
110
+
111
+ <!-- This section describes the evaluation protocols and provides the results. -->
112
+
113
+ ### Testing Data, Factors & Metrics
114
+
115
+ #### Testing Data
116
+
117
+ <!-- This should link to a Dataset Card if possible. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Factors
122
+
123
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
124
+
125
+ [More Information Needed]
126
+
127
+ #### Metrics
128
+
129
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
130
+
131
+ [More Information Needed]
132
+
133
+ ### Results
134
+
135
+ [More Information Needed]
136
+
137
+ #### Summary
138
+
139
+
140
+
141
+ ## Model Examination [optional]
142
+
143
+ <!-- Relevant interpretability work for the model goes here -->
144
+
145
+ [More Information Needed]
146
+
147
+ ## Environmental Impact
148
+
149
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
150
+
151
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
152
+
153
+ - **Hardware Type:** [More Information Needed]
154
+ - **Hours used:** [More Information Needed]
155
+ - **Cloud Provider:** [More Information Needed]
156
+ - **Compute Region:** [More Information Needed]
157
+ - **Carbon Emitted:** [More Information Needed]
158
+
159
+ ## Technical Specifications [optional]
160
+
161
+ ### Model Architecture and Objective
162
+
163
+ [More Information Needed]
164
+
165
+ ### Compute Infrastructure
166
+
167
+ [More Information Needed]
168
+
169
+ #### Hardware
170
+
171
+ [More Information Needed]
172
+
173
+ #### Software
174
+
175
+ [More Information Needed]
176
+
177
+ ## Citation [optional]
178
+
179
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
180
+
181
+ **BibTeX:**
182
+
183
+ [More Information Needed]
184
+
185
+ **APA:**
186
+
187
+ [More Information Needed]
188
+
189
+ ## Glossary [optional]
190
+
191
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
192
+
193
+ [More Information Needed]
194
+
195
+ ## More Information [optional]
196
+
197
+ [More Information Needed]
198
+
199
+ ## Model Card Authors [optional]
200
+
201
+ [More Information Needed]
202
+
203
+ ## Model Card Contact
204
+
205
+ [More Information Needed]
206
+ ### Framework versions
207
+
208
+ - PEFT 0.17.0
checkpoint-315/adapter_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen3-4B-Instruct-2507",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": null,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 8,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "q_proj",
29
+ "gate_proj",
30
+ "up_proj",
31
+ "v_proj",
32
+ "down_proj",
33
+ "k_proj",
34
+ "o_proj"
35
+ ],
36
+ "target_parameters": [],
37
+ "task_type": "CAUSAL_LM",
38
+ "trainable_token_indices": null,
39
+ "use_dora": false,
40
+ "use_qalora": false,
41
+ "use_rslora": false
42
+ }
checkpoint-315/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e45b31c137803c2351f0144a9d9cdee51846c1e3ed927ec07b20c1cc38ea3b8
3
+ size 66126768
checkpoint-315/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
checkpoint-315/chat_template.jinja ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if message.content is string %}
27
+ {%- set content = message.content %}
28
+ {%- else %}
29
+ {%- set content = '' %}
30
+ {%- endif %}
31
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
+ {%- elif message.role == "assistant" %}
34
+ {%- set reasoning_content = '' %}
35
+ {%- if message.reasoning_content is string %}
36
+ {%- set reasoning_content = message.reasoning_content %}
37
+ {%- else %}
38
+ {%- if '</think>' in content %}
39
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
+ {%- endif %}
42
+ {%- endif %}
43
+ {%- if loop.index0 > ns.last_query_index %}
44
+ {%- if loop.last or (not loop.last and reasoning_content) %}
45
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
+ {%- else %}
47
+ {{- '<|im_start|>' + message.role + '\n' + content }}
48
+ {%- endif %}
49
+ {%- else %}
50
+ {{- '<|im_start|>' + message.role + '\n' + content }}
51
+ {%- endif %}
52
+ {%- if message.tool_calls %}
53
+ {%- for tool_call in message.tool_calls %}
54
+ {%- if (loop.first and content) or (not loop.first) %}
55
+ {{- '\n' }}
56
+ {%- endif %}
57
+ {%- if tool_call.function %}
58
+ {%- set tool_call = tool_call.function %}
59
+ {%- endif %}
60
+ {{- '<tool_call>\n{"name": "' }}
61
+ {{- tool_call.name }}
62
+ {{- '", "arguments": ' }}
63
+ {%- if tool_call.arguments is string %}
64
+ {{- tool_call.arguments }}
65
+ {%- else %}
66
+ {{- tool_call.arguments | tojson }}
67
+ {%- endif %}
68
+ {{- '}\n</tool_call>' }}
69
+ {%- endfor %}
70
+ {%- endif %}
71
+ {{- '<|im_end|>\n' }}
72
+ {%- elif message.role == "tool" %}
73
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
74
+ {{- '<|im_start|>user' }}
75
+ {%- endif %}
76
+ {{- '\n<tool_response>\n' }}
77
+ {{- content }}
78
+ {{- '\n</tool_response>' }}
79
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
80
+ {{- '<|im_end|>\n' }}
81
+ {%- endif %}
82
+ {%- endif %}
83
+ {%- endfor %}
84
+ {%- if add_generation_prompt %}
85
+ {{- '<|im_start|>assistant\n' }}
86
+ {%- endif %}
checkpoint-315/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-315/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d44680f2fdfc4ac7c7dd0e7295df76182dc0e40fb6ac329f5cbc408263404a20
3
+ size 34136372
checkpoint-315/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2d7d2a5d0cd064a913774d809a2a7705ac965a3966e312d27f543b5be7db2a2
3
+ size 14512
checkpoint-315/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee6d6b8378064132b3fefd1a59db9b8c5cbbd76ac6045dd4c4cc3f89819a94f9
3
+ size 14512
checkpoint-315/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ae1d9962c79cee29f1198d64dd25fad600fe6cac4cd38a92b31a92040718648
3
+ size 1064
checkpoint-315/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
checkpoint-315/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654