yifeigao commited on
Commit
9d334a0
·
verified ·
1 Parent(s): a2f7acb

Upload folder using huggingface_hub

Browse files
Files changed (45) hide show
  1. .DS_Store +0 -0
  2. .gitattributes +3 -0
  3. TextUI-Cap-7B/README.md +202 -0
  4. TextUI-Cap-7B/adapter_config.json +39 -0
  5. TextUI-Cap-7B/adapter_model.safetensors +3 -0
  6. TextUI-Cap-7B/added_tokens.json +24 -0
  7. TextUI-Cap-7B/merges.txt +0 -0
  8. TextUI-Cap-7B/special_tokens_map.json +31 -0
  9. TextUI-Cap-7B/tokenizer.json +3 -0
  10. TextUI-Cap-7B/tokenizer_config.json +209 -0
  11. TextUI-Cap-7B/trainer_log.jsonl +110 -0
  12. TextUI-Cap-7B/trainer_state.json +814 -0
  13. TextUI-Cap-7B/training_args.bin +3 -0
  14. TextUI-Cap-7B/vocab.json +0 -0
  15. TextUI-Func-7B/README.md +76 -0
  16. TextUI-Func-7B/adapter_config.json +39 -0
  17. TextUI-Func-7B/adapter_model.safetensors +3 -0
  18. TextUI-Func-7B/added_tokens.json +24 -0
  19. TextUI-Func-7B/all_results.json +12 -0
  20. TextUI-Func-7B/eval_results.json +7 -0
  21. TextUI-Func-7B/merges.txt +0 -0
  22. TextUI-Func-7B/runs/events.out.tfevents.1749786583.autodl-container-40b74f9912-ad7efc16.23496.0 +3 -0
  23. TextUI-Func-7B/runs/events.out.tfevents.1749796231.autodl-container-40b74f9912-ad7efc16.23496.1 +3 -0
  24. TextUI-Func-7B/special_tokens_map.json +31 -0
  25. TextUI-Func-7B/tokenizer.json +3 -0
  26. TextUI-Func-7B/tokenizer_config.json +209 -0
  27. TextUI-Func-7B/train_results.json +8 -0
  28. TextUI-Func-7B/trainer_log.jsonl +261 -0
  29. TextUI-Func-7B/trainer_state.json +1873 -0
  30. TextUI-Func-7B/training_args.bin +3 -0
  31. TextUI-Func-7B/training_eval_loss.png +0 -0
  32. TextUI-Func-7B/training_loss.png +0 -0
  33. TextUI-Func-7B/vocab.json +0 -0
  34. TextUI-Trans-7B/README.md +202 -0
  35. TextUI-Trans-7B/adapter_config.json +39 -0
  36. TextUI-Trans-7B/adapter_model.safetensors +3 -0
  37. TextUI-Trans-7B/added_tokens.json +24 -0
  38. TextUI-Trans-7B/merges.txt +0 -0
  39. TextUI-Trans-7B/special_tokens_map.json +31 -0
  40. TextUI-Trans-7B/tokenizer.json +3 -0
  41. TextUI-Trans-7B/tokenizer_config.json +209 -0
  42. TextUI-Trans-7B/trainer_log.jsonl +169 -0
  43. TextUI-Trans-7B/trainer_state.json +1132 -0
  44. TextUI-Trans-7B/training_args.bin +3 -0
  45. TextUI-Trans-7B/vocab.json +0 -0
.DS_Store ADDED
Binary file (10.2 kB). View file
 
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ TextUI-Cap-7B/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ TextUI-Func-7B/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ TextUI-Trans-7B/tokenizer.json filter=lfs diff=lfs merge=lfs -text
TextUI-Cap-7B/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /root/autodl-fs/Qwen2.5-7B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.1
TextUI-Cap-7B/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/root/autodl-fs/Qwen2.5-7B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "v_proj",
28
+ "q_proj",
29
+ "up_proj",
30
+ "gate_proj",
31
+ "down_proj",
32
+ "k_proj",
33
+ "o_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
TextUI-Cap-7B/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e1ffe3e8171c896b957a27c4245fa27db8abdb734239414307fb8f345d49211
3
+ size 161533192
TextUI-Cap-7B/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
TextUI-Cap-7B/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
TextUI-Cap-7B/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
TextUI-Cap-7B/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
TextUI-Cap-7B/tokenizer_config.json ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "right",
206
+ "split_special_tokens": false,
207
+ "tokenizer_class": "Qwen2Tokenizer",
208
+ "unk_token": null
209
+ }
TextUI-Cap-7B/trainer_log.jsonl ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"current_steps": 1, "total_steps": 100, "loss": 1.9336, "lr": 0.0, "epoch": 0.03524229074889868, "percentage": 1.0, "elapsed_time": "0:00:15", "remaining_time": "0:25:05"}
2
+ {"current_steps": 2, "total_steps": 100, "loss": 2.048, "lr": 1e-05, "epoch": 0.07048458149779736, "percentage": 2.0, "elapsed_time": "0:00:33", "remaining_time": "0:27:38"}
3
+ {"current_steps": 3, "total_steps": 100, "loss": 1.9311, "lr": 2e-05, "epoch": 0.10572687224669604, "percentage": 3.0, "elapsed_time": "0:00:56", "remaining_time": "0:30:26"}
4
+ {"current_steps": 4, "total_steps": 100, "loss": 1.9814, "lr": 3e-05, "epoch": 0.14096916299559473, "percentage": 4.0, "elapsed_time": "0:01:16", "remaining_time": "0:30:33"}
5
+ {"current_steps": 5, "total_steps": 100, "loss": 1.9955, "lr": 4e-05, "epoch": 0.1762114537444934, "percentage": 5.0, "elapsed_time": "0:01:32", "remaining_time": "0:29:14"}
6
+ {"current_steps": 6, "total_steps": 100, "loss": 1.874, "lr": 5e-05, "epoch": 0.21145374449339208, "percentage": 6.0, "elapsed_time": "0:01:48", "remaining_time": "0:28:18"}
7
+ {"current_steps": 7, "total_steps": 100, "loss": 1.7726, "lr": 6e-05, "epoch": 0.24669603524229075, "percentage": 7.0, "elapsed_time": "0:02:09", "remaining_time": "0:28:44"}
8
+ {"current_steps": 8, "total_steps": 100, "loss": 1.6056, "lr": 7e-05, "epoch": 0.28193832599118945, "percentage": 8.0, "elapsed_time": "0:02:26", "remaining_time": "0:28:04"}
9
+ {"current_steps": 9, "total_steps": 100, "loss": 1.4883, "lr": 8e-05, "epoch": 0.31718061674008813, "percentage": 9.0, "elapsed_time": "0:02:44", "remaining_time": "0:27:45"}
10
+ {"current_steps": 10, "total_steps": 100, "loss": 1.3543, "lr": 9e-05, "epoch": 0.3524229074889868, "percentage": 10.0, "elapsed_time": "0:03:07", "remaining_time": "0:28:07"}
11
+ {"current_steps": 10, "total_steps": 100, "eval_loss": 1.2461069822311401, "epoch": 0.3524229074889868, "percentage": 10.0, "elapsed_time": "0:03:32", "remaining_time": "0:31:55"}
12
+ {"current_steps": 11, "total_steps": 100, "loss": 1.253, "lr": 0.0001, "epoch": 0.3876651982378855, "percentage": 11.0, "elapsed_time": "0:03:48", "remaining_time": "0:30:49"}
13
+ {"current_steps": 12, "total_steps": 100, "loss": 1.1504, "lr": 0.00011000000000000002, "epoch": 0.42290748898678415, "percentage": 12.0, "elapsed_time": "0:04:05", "remaining_time": "0:29:58"}
14
+ {"current_steps": 13, "total_steps": 100, "loss": 1.1904, "lr": 0.00012, "epoch": 0.4581497797356828, "percentage": 13.0, "elapsed_time": "0:04:23", "remaining_time": "0:29:25"}
15
+ {"current_steps": 14, "total_steps": 100, "loss": 1.065, "lr": 0.00013000000000000002, "epoch": 0.4933920704845815, "percentage": 14.0, "elapsed_time": "0:04:40", "remaining_time": "0:28:42"}
16
+ {"current_steps": 15, "total_steps": 100, "loss": 1.0323, "lr": 0.00014, "epoch": 0.5286343612334802, "percentage": 15.0, "elapsed_time": "0:05:01", "remaining_time": "0:28:26"}
17
+ {"current_steps": 16, "total_steps": 100, "loss": 0.967, "lr": 0.00015000000000000001, "epoch": 0.5638766519823789, "percentage": 16.0, "elapsed_time": "0:05:16", "remaining_time": "0:27:43"}
18
+ {"current_steps": 17, "total_steps": 100, "loss": 0.9886, "lr": 0.00016, "epoch": 0.5991189427312775, "percentage": 17.0, "elapsed_time": "0:05:34", "remaining_time": "0:27:12"}
19
+ {"current_steps": 18, "total_steps": 100, "loss": 0.9052, "lr": 0.00017, "epoch": 0.6343612334801763, "percentage": 18.0, "elapsed_time": "0:05:51", "remaining_time": "0:26:39"}
20
+ {"current_steps": 19, "total_steps": 100, "loss": 0.8916, "lr": 0.00018, "epoch": 0.6696035242290749, "percentage": 19.0, "elapsed_time": "0:06:12", "remaining_time": "0:26:26"}
21
+ {"current_steps": 20, "total_steps": 100, "loss": 0.8508, "lr": 0.00019, "epoch": 0.7048458149779736, "percentage": 20.0, "elapsed_time": "0:06:29", "remaining_time": "0:25:57"}
22
+ {"current_steps": 20, "total_steps": 100, "eval_loss": 0.8256819844245911, "epoch": 0.7048458149779736, "percentage": 20.0, "elapsed_time": "0:06:54", "remaining_time": "0:27:39"}
23
+ {"current_steps": 21, "total_steps": 100, "loss": 0.837, "lr": 0.0002, "epoch": 0.7400881057268722, "percentage": 21.0, "elapsed_time": "0:07:10", "remaining_time": "0:27:00"}
24
+ {"current_steps": 22, "total_steps": 100, "loss": 0.8872, "lr": 0.0001999229036240723, "epoch": 0.775330396475771, "percentage": 22.0, "elapsed_time": "0:07:30", "remaining_time": "0:26:35"}
25
+ {"current_steps": 23, "total_steps": 100, "loss": 0.8076, "lr": 0.0001996917333733128, "epoch": 0.8105726872246696, "percentage": 23.0, "elapsed_time": "0:07:47", "remaining_time": "0:26:04"}
26
+ {"current_steps": 24, "total_steps": 100, "loss": 0.8181, "lr": 0.00019930684569549264, "epoch": 0.8458149779735683, "percentage": 24.0, "elapsed_time": "0:08:02", "remaining_time": "0:25:28"}
27
+ {"current_steps": 25, "total_steps": 100, "loss": 0.8081, "lr": 0.00019876883405951377, "epoch": 0.8810572687224669, "percentage": 25.0, "elapsed_time": "0:08:20", "remaining_time": "0:25:01"}
28
+ {"current_steps": 26, "total_steps": 100, "loss": 0.7206, "lr": 0.00019807852804032305, "epoch": 0.9162995594713657, "percentage": 26.0, "elapsed_time": "0:08:34", "remaining_time": "0:24:23"}
29
+ {"current_steps": 27, "total_steps": 100, "loss": 0.7821, "lr": 0.00019723699203976766, "epoch": 0.9515418502202643, "percentage": 27.0, "elapsed_time": "0:08:52", "remaining_time": "0:23:58"}
30
+ {"current_steps": 28, "total_steps": 100, "loss": 0.7356, "lr": 0.00019624552364536473, "epoch": 0.986784140969163, "percentage": 28.0, "elapsed_time": "0:09:10", "remaining_time": "0:23:34"}
31
+ {"current_steps": 29, "total_steps": 100, "loss": 0.7774, "lr": 0.00019510565162951537, "epoch": 1.0, "percentage": 29.0, "elapsed_time": "0:09:20", "remaining_time": "0:22:53"}
32
+ {"current_steps": 30, "total_steps": 100, "loss": 0.7069, "lr": 0.00019381913359224842, "epoch": 1.0352422907488987, "percentage": 30.0, "elapsed_time": "0:09:35", "remaining_time": "0:22:22"}
33
+ {"current_steps": 30, "total_steps": 100, "eval_loss": 0.6948859691619873, "epoch": 1.0352422907488987, "percentage": 30.0, "elapsed_time": "0:10:00", "remaining_time": "0:23:21"}
34
+ {"current_steps": 31, "total_steps": 100, "loss": 0.7228, "lr": 0.0001923879532511287, "epoch": 1.0704845814977975, "percentage": 31.0, "elapsed_time": "0:10:19", "remaining_time": "0:22:58"}
35
+ {"current_steps": 32, "total_steps": 100, "loss": 0.6692, "lr": 0.00019081431738250814, "epoch": 1.105726872246696, "percentage": 32.0, "elapsed_time": "0:10:36", "remaining_time": "0:22:33"}
36
+ {"current_steps": 33, "total_steps": 100, "loss": 0.6742, "lr": 0.0001891006524188368, "epoch": 1.1409691629955947, "percentage": 33.0, "elapsed_time": "0:10:52", "remaining_time": "0:22:05"}
37
+ {"current_steps": 34, "total_steps": 100, "loss": 0.7012, "lr": 0.00018724960070727972, "epoch": 1.1762114537444934, "percentage": 34.0, "elapsed_time": "0:11:07", "remaining_time": "0:21:36"}
38
+ {"current_steps": 35, "total_steps": 100, "loss": 0.6594, "lr": 0.00018526401643540922, "epoch": 1.2114537444933922, "percentage": 35.0, "elapsed_time": "0:11:25", "remaining_time": "0:21:13"}
39
+ {"current_steps": 36, "total_steps": 100, "loss": 0.6723, "lr": 0.00018314696123025454, "epoch": 1.2466960352422907, "percentage": 36.0, "elapsed_time": "0:11:41", "remaining_time": "0:20:46"}
40
+ {"current_steps": 37, "total_steps": 100, "loss": 0.6514, "lr": 0.00018090169943749476, "epoch": 1.2819383259911894, "percentage": 37.0, "elapsed_time": "0:11:57", "remaining_time": "0:20:21"}
41
+ {"current_steps": 38, "total_steps": 100, "loss": 0.6585, "lr": 0.00017853169308807448, "epoch": 1.3171806167400881, "percentage": 38.0, "elapsed_time": "0:12:21", "remaining_time": "0:20:09"}
42
+ {"current_steps": 39, "total_steps": 100, "loss": 0.6339, "lr": 0.0001760405965600031, "epoch": 1.3524229074889869, "percentage": 39.0, "elapsed_time": "0:12:37", "remaining_time": "0:19:44"}
43
+ {"current_steps": 40, "total_steps": 100, "loss": 0.655, "lr": 0.00017343225094356855, "epoch": 1.3876651982378854, "percentage": 40.0, "elapsed_time": "0:12:52", "remaining_time": "0:19:18"}
44
+ {"current_steps": 40, "total_steps": 100, "eval_loss": 0.6199031472206116, "epoch": 1.3876651982378854, "percentage": 40.0, "elapsed_time": "0:13:17", "remaining_time": "0:19:56"}
45
+ {"current_steps": 41, "total_steps": 100, "loss": 0.6334, "lr": 0.00017071067811865476, "epoch": 1.422907488986784, "percentage": 41.0, "elapsed_time": "0:13:37", "remaining_time": "0:19:35"}
46
+ {"current_steps": 42, "total_steps": 100, "loss": 0.644, "lr": 0.0001678800745532942, "epoch": 1.4581497797356828, "percentage": 42.0, "elapsed_time": "0:13:52", "remaining_time": "0:19:09"}
47
+ {"current_steps": 43, "total_steps": 100, "loss": 0.5925, "lr": 0.00016494480483301836, "epoch": 1.4933920704845816, "percentage": 43.0, "elapsed_time": "0:14:05", "remaining_time": "0:18:41"}
48
+ {"current_steps": 44, "total_steps": 100, "loss": 0.6486, "lr": 0.00016190939493098344, "epoch": 1.5286343612334803, "percentage": 44.0, "elapsed_time": "0:14:22", "remaining_time": "0:18:17"}
49
+ {"current_steps": 45, "total_steps": 100, "loss": 0.6039, "lr": 0.00015877852522924732, "epoch": 1.5638766519823788, "percentage": 45.0, "elapsed_time": "0:14:38", "remaining_time": "0:17:53"}
50
+ {"current_steps": 46, "total_steps": 100, "loss": 0.6244, "lr": 0.00015555702330196023, "epoch": 1.5991189427312775, "percentage": 46.0, "elapsed_time": "0:15:02", "remaining_time": "0:17:39"}
51
+ {"current_steps": 47, "total_steps": 100, "loss": 0.5672, "lr": 0.0001522498564715949, "epoch": 1.6343612334801763, "percentage": 47.0, "elapsed_time": "0:15:18", "remaining_time": "0:17:15"}
52
+ {"current_steps": 48, "total_steps": 100, "loss": 0.6356, "lr": 0.00014886212414969553, "epoch": 1.6696035242290748, "percentage": 48.0, "elapsed_time": "0:15:36", "remaining_time": "0:16:54"}
53
+ {"current_steps": 49, "total_steps": 100, "loss": 0.6105, "lr": 0.00014539904997395468, "epoch": 1.7048458149779737, "percentage": 49.0, "elapsed_time": "0:15:56", "remaining_time": "0:16:35"}
54
+ {"current_steps": 50, "total_steps": 100, "loss": 0.6023, "lr": 0.0001418659737537428, "epoch": 1.7400881057268722, "percentage": 50.0, "elapsed_time": "0:16:18", "remaining_time": "0:16:18"}
55
+ {"current_steps": 50, "total_steps": 100, "eval_loss": 0.5753047466278076, "epoch": 1.7400881057268722, "percentage": 50.0, "elapsed_time": "0:16:43", "remaining_time": "0:16:43"}
56
+ {"current_steps": 51, "total_steps": 100, "loss": 0.6216, "lr": 0.000138268343236509, "epoch": 1.775330396475771, "percentage": 51.0, "elapsed_time": "0:17:03", "remaining_time": "0:16:23"}
57
+ {"current_steps": 52, "total_steps": 100, "loss": 0.6124, "lr": 0.0001346117057077493, "epoch": 1.8105726872246697, "percentage": 52.0, "elapsed_time": "0:17:19", "remaining_time": "0:15:59"}
58
+ {"current_steps": 53, "total_steps": 100, "loss": 0.5445, "lr": 0.00013090169943749476, "epoch": 1.8458149779735682, "percentage": 53.0, "elapsed_time": "0:17:39", "remaining_time": "0:15:39"}
59
+ {"current_steps": 54, "total_steps": 100, "loss": 0.5508, "lr": 0.00012714404498650743, "epoch": 1.881057268722467, "percentage": 54.0, "elapsed_time": "0:17:55", "remaining_time": "0:15:16"}
60
+ {"current_steps": 55, "total_steps": 100, "loss": 0.5522, "lr": 0.00012334453638559057, "epoch": 1.9162995594713657, "percentage": 55.0, "elapsed_time": "0:18:13", "remaining_time": "0:14:54"}
61
+ {"current_steps": 56, "total_steps": 100, "loss": 0.5569, "lr": 0.00011950903220161285, "epoch": 1.9515418502202642, "percentage": 56.0, "elapsed_time": "0:18:29", "remaining_time": "0:14:31"}
62
+ {"current_steps": 57, "total_steps": 100, "loss": 0.5307, "lr": 0.0001156434465040231, "epoch": 1.9867841409691631, "percentage": 57.0, "elapsed_time": "0:18:51", "remaining_time": "0:14:13"}
63
+ {"current_steps": 58, "total_steps": 100, "loss": 0.6131, "lr": 0.00011175373974578378, "epoch": 2.0, "percentage": 58.0, "elapsed_time": "0:18:58", "remaining_time": "0:13:44"}
64
+ {"current_steps": 59, "total_steps": 100, "loss": 0.4965, "lr": 0.0001078459095727845, "epoch": 2.0352422907488985, "percentage": 59.0, "elapsed_time": "0:19:14", "remaining_time": "0:13:22"}
65
+ {"current_steps": 60, "total_steps": 100, "loss": 0.5086, "lr": 0.00010392598157590688, "epoch": 2.0704845814977975, "percentage": 60.0, "elapsed_time": "0:19:31", "remaining_time": "0:13:00"}
66
+ {"current_steps": 60, "total_steps": 100, "eval_loss": 0.5449034571647644, "epoch": 2.0704845814977975, "percentage": 60.0, "elapsed_time": "0:19:56", "remaining_time": "0:13:17"}
67
+ {"current_steps": 61, "total_steps": 100, "loss": 0.5208, "lr": 0.0001, "epoch": 2.105726872246696, "percentage": 61.0, "elapsed_time": "0:20:13", "remaining_time": "0:12:55"}
68
+ {"current_steps": 62, "total_steps": 100, "loss": 0.5298, "lr": 9.607401842409317e-05, "epoch": 2.140969162995595, "percentage": 62.0, "elapsed_time": "0:20:33", "remaining_time": "0:12:36"}
69
+ {"current_steps": 63, "total_steps": 100, "loss": 0.5025, "lr": 9.215409042721552e-05, "epoch": 2.1762114537444934, "percentage": 63.0, "elapsed_time": "0:20:53", "remaining_time": "0:12:16"}
70
+ {"current_steps": 64, "total_steps": 100, "loss": 0.4856, "lr": 8.824626025421626e-05, "epoch": 2.211453744493392, "percentage": 64.0, "elapsed_time": "0:21:08", "remaining_time": "0:11:53"}
71
+ {"current_steps": 65, "total_steps": 100, "loss": 0.5118, "lr": 8.435655349597689e-05, "epoch": 2.246696035242291, "percentage": 65.0, "elapsed_time": "0:21:26", "remaining_time": "0:11:32"}
72
+ {"current_steps": 66, "total_steps": 100, "loss": 0.5256, "lr": 8.049096779838719e-05, "epoch": 2.2819383259911894, "percentage": 66.0, "elapsed_time": "0:21:42", "remaining_time": "0:11:10"}
73
+ {"current_steps": 67, "total_steps": 100, "loss": 0.5396, "lr": 7.66554636144095e-05, "epoch": 2.317180616740088, "percentage": 67.0, "elapsed_time": "0:22:03", "remaining_time": "0:10:52"}
74
+ {"current_steps": 68, "total_steps": 100, "loss": 0.5633, "lr": 7.285595501349258e-05, "epoch": 2.352422907488987, "percentage": 68.0, "elapsed_time": "0:22:27", "remaining_time": "0:10:34"}
75
+ {"current_steps": 69, "total_steps": 100, "loss": 0.5233, "lr": 6.909830056250527e-05, "epoch": 2.3876651982378854, "percentage": 69.0, "elapsed_time": "0:22:41", "remaining_time": "0:10:11"}
76
+ {"current_steps": 70, "total_steps": 100, "loss": 0.5638, "lr": 6.538829429225069e-05, "epoch": 2.4229074889867843, "percentage": 70.0, "elapsed_time": "0:22:58", "remaining_time": "0:09:50"}
77
+ {"current_steps": 70, "total_steps": 100, "eval_loss": 0.5274443030357361, "epoch": 2.4229074889867843, "percentage": 70.0, "elapsed_time": "0:23:23", "remaining_time": "0:10:01"}
78
+ {"current_steps": 71, "total_steps": 100, "loss": 0.5003, "lr": 6.173165676349103e-05, "epoch": 2.458149779735683, "percentage": 71.0, "elapsed_time": "0:23:41", "remaining_time": "0:09:40"}
79
+ {"current_steps": 72, "total_steps": 100, "loss": 0.4974, "lr": 5.8134026246257225e-05, "epoch": 2.4933920704845813, "percentage": 72.0, "elapsed_time": "0:23:56", "remaining_time": "0:09:18"}
80
+ {"current_steps": 73, "total_steps": 100, "loss": 0.5325, "lr": 5.4600950026045326e-05, "epoch": 2.5286343612334803, "percentage": 73.0, "elapsed_time": "0:24:17", "remaining_time": "0:08:59"}
81
+ {"current_steps": 74, "total_steps": 100, "loss": 0.4323, "lr": 5.113787585030454e-05, "epoch": 2.563876651982379, "percentage": 74.0, "elapsed_time": "0:24:36", "remaining_time": "0:08:38"}
82
+ {"current_steps": 75, "total_steps": 100, "loss": 0.4905, "lr": 4.7750143528405126e-05, "epoch": 2.5991189427312777, "percentage": 75.0, "elapsed_time": "0:24:52", "remaining_time": "0:08:17"}
83
+ {"current_steps": 76, "total_steps": 100, "loss": 0.4649, "lr": 4.444297669803981e-05, "epoch": 2.6343612334801763, "percentage": 76.0, "elapsed_time": "0:25:08", "remaining_time": "0:07:56"}
84
+ {"current_steps": 77, "total_steps": 100, "loss": 0.5379, "lr": 4.12214747707527e-05, "epoch": 2.6696035242290748, "percentage": 77.0, "elapsed_time": "0:25:29", "remaining_time": "0:07:36"}
85
+ {"current_steps": 78, "total_steps": 100, "loss": 0.4885, "lr": 3.8090605069016595e-05, "epoch": 2.7048458149779737, "percentage": 78.0, "elapsed_time": "0:25:49", "remaining_time": "0:07:16"}
86
+ {"current_steps": 79, "total_steps": 100, "loss": 0.4923, "lr": 3.5055195166981645e-05, "epoch": 2.7400881057268722, "percentage": 79.0, "elapsed_time": "0:26:09", "remaining_time": "0:06:57"}
87
+ {"current_steps": 80, "total_steps": 100, "loss": 0.4747, "lr": 3.211992544670582e-05, "epoch": 2.7753303964757707, "percentage": 80.0, "elapsed_time": "0:26:27", "remaining_time": "0:06:36"}
88
+ {"current_steps": 80, "total_steps": 100, "eval_loss": 0.5167238712310791, "epoch": 2.7753303964757707, "percentage": 80.0, "elapsed_time": "0:26:52", "remaining_time": "0:06:43"}
89
+ {"current_steps": 81, "total_steps": 100, "loss": 0.4887, "lr": 2.9289321881345254e-05, "epoch": 2.8105726872246697, "percentage": 81.0, "elapsed_time": "0:27:11", "remaining_time": "0:06:22"}
90
+ {"current_steps": 82, "total_steps": 100, "loss": 0.4812, "lr": 2.6567749056431467e-05, "epoch": 2.845814977973568, "percentage": 82.0, "elapsed_time": "0:27:32", "remaining_time": "0:06:02"}
91
+ {"current_steps": 83, "total_steps": 100, "loss": 0.4947, "lr": 2.3959403439996907e-05, "epoch": 2.8810572687224667, "percentage": 83.0, "elapsed_time": "0:27:49", "remaining_time": "0:05:41"}
92
+ {"current_steps": 84, "total_steps": 100, "loss": 0.5061, "lr": 2.146830691192553e-05, "epoch": 2.9162995594713657, "percentage": 84.0, "elapsed_time": "0:28:08", "remaining_time": "0:05:21"}
93
+ {"current_steps": 85, "total_steps": 100, "loss": 0.4976, "lr": 1.9098300562505266e-05, "epoch": 2.951541850220264, "percentage": 85.0, "elapsed_time": "0:28:24", "remaining_time": "0:05:00"}
94
+ {"current_steps": 86, "total_steps": 100, "loss": 0.459, "lr": 1.6853038769745467e-05, "epoch": 2.986784140969163, "percentage": 86.0, "elapsed_time": "0:28:43", "remaining_time": "0:04:40"}
95
+ {"current_steps": 87, "total_steps": 100, "loss": 0.4275, "lr": 1.4735983564590783e-05, "epoch": 3.0, "percentage": 87.0, "elapsed_time": "0:28:50", "remaining_time": "0:04:18"}
96
+ {"current_steps": 88, "total_steps": 100, "loss": 0.4549, "lr": 1.2750399292720283e-05, "epoch": 3.0352422907488985, "percentage": 88.0, "elapsed_time": "0:29:10", "remaining_time": "0:03:58"}
97
+ {"current_steps": 89, "total_steps": 100, "loss": 0.4918, "lr": 1.0899347581163221e-05, "epoch": 3.0704845814977975, "percentage": 89.0, "elapsed_time": "0:29:29", "remaining_time": "0:03:38"}
98
+ {"current_steps": 90, "total_steps": 100, "loss": 0.458, "lr": 9.185682617491863e-06, "epoch": 3.105726872246696, "percentage": 90.0, "elapsed_time": "0:29:54", "remaining_time": "0:03:19"}
99
+ {"current_steps": 90, "total_steps": 100, "eval_loss": 0.5118579268455505, "epoch": 3.105726872246696, "percentage": 90.0, "elapsed_time": "0:30:19", "remaining_time": "0:03:22"}
100
+ {"current_steps": 91, "total_steps": 100, "loss": 0.4861, "lr": 7.612046748871327e-06, "epoch": 3.140969162995595, "percentage": 91.0, "elapsed_time": "0:30:40", "remaining_time": "0:03:02"}
101
+ {"current_steps": 92, "total_steps": 100, "loss": 0.4881, "lr": 6.180866407751595e-06, "epoch": 3.1762114537444934, "percentage": 92.0, "elapsed_time": "0:30:58", "remaining_time": "0:02:41"}
102
+ {"current_steps": 93, "total_steps": 100, "loss": 0.4524, "lr": 4.8943483704846475e-06, "epoch": 3.211453744493392, "percentage": 93.0, "elapsed_time": "0:31:14", "remaining_time": "0:02:21"}
103
+ {"current_steps": 94, "total_steps": 100, "loss": 0.4678, "lr": 3.7544763546352834e-06, "epoch": 3.246696035242291, "percentage": 94.0, "elapsed_time": "0:31:31", "remaining_time": "0:02:00"}
104
+ {"current_steps": 95, "total_steps": 100, "loss": 0.5017, "lr": 2.7630079602323442e-06, "epoch": 3.2819383259911894, "percentage": 95.0, "elapsed_time": "0:31:47", "remaining_time": "0:01:40"}
105
+ {"current_steps": 96, "total_steps": 100, "loss": 0.4615, "lr": 1.921471959676957e-06, "epoch": 3.317180616740088, "percentage": 96.0, "elapsed_time": "0:32:02", "remaining_time": "0:01:20"}
106
+ {"current_steps": 97, "total_steps": 100, "loss": 0.4972, "lr": 1.231165940486234e-06, "epoch": 3.352422907488987, "percentage": 97.0, "elapsed_time": "0:32:20", "remaining_time": "0:01:00"}
107
+ {"current_steps": 98, "total_steps": 100, "loss": 0.4188, "lr": 6.931543045073708e-07, "epoch": 3.3876651982378854, "percentage": 98.0, "elapsed_time": "0:32:37", "remaining_time": "0:00:39"}
108
+ {"current_steps": 99, "total_steps": 100, "loss": 0.4577, "lr": 3.0826662668720364e-07, "epoch": 3.4229074889867843, "percentage": 99.0, "elapsed_time": "0:32:54", "remaining_time": "0:00:19"}
109
+ {"current_steps": 100, "total_steps": 100, "loss": 0.4505, "lr": 7.709637592770991e-08, "epoch": 3.458149779735683, "percentage": 100.0, "elapsed_time": "0:33:15", "remaining_time": "0:00:00"}
110
+ {"current_steps": 100, "total_steps": 100, "eval_loss": 0.5112941265106201, "epoch": 3.458149779735683, "percentage": 100.0, "elapsed_time": "0:33:41", "remaining_time": "0:00:00"}
TextUI-Cap-7B/trainer_state.json ADDED
@@ -0,0 +1,814 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 100,
3
+ "best_metric": 0.5112941265106201,
4
+ "best_model_checkpoint": "/root/autodl-tmp/model/lora-textui/stage0_caption-v2/checkpoint-100",
5
+ "epoch": 3.458149779735683,
6
+ "eval_steps": 10,
7
+ "global_step": 100,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.03524229074889868,
14
+ "grad_norm": 1.2463780641555786,
15
+ "learning_rate": 0.0,
16
+ "loss": 1.9336,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.07048458149779736,
21
+ "grad_norm": 1.3275448083877563,
22
+ "learning_rate": 1e-05,
23
+ "loss": 2.048,
24
+ "step": 2
25
+ },
26
+ {
27
+ "epoch": 0.10572687224669604,
28
+ "grad_norm": 1.2619495391845703,
29
+ "learning_rate": 2e-05,
30
+ "loss": 1.9311,
31
+ "step": 3
32
+ },
33
+ {
34
+ "epoch": 0.14096916299559473,
35
+ "grad_norm": 1.2719473838806152,
36
+ "learning_rate": 3e-05,
37
+ "loss": 1.9814,
38
+ "step": 4
39
+ },
40
+ {
41
+ "epoch": 0.1762114537444934,
42
+ "grad_norm": 1.3939034938812256,
43
+ "learning_rate": 4e-05,
44
+ "loss": 1.9955,
45
+ "step": 5
46
+ },
47
+ {
48
+ "epoch": 0.21145374449339208,
49
+ "grad_norm": 1.3475849628448486,
50
+ "learning_rate": 5e-05,
51
+ "loss": 1.874,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 0.24669603524229075,
56
+ "grad_norm": 1.2599456310272217,
57
+ "learning_rate": 6e-05,
58
+ "loss": 1.7726,
59
+ "step": 7
60
+ },
61
+ {
62
+ "epoch": 0.28193832599118945,
63
+ "grad_norm": 1.1231255531311035,
64
+ "learning_rate": 7e-05,
65
+ "loss": 1.6056,
66
+ "step": 8
67
+ },
68
+ {
69
+ "epoch": 0.31718061674008813,
70
+ "grad_norm": 0.926201581954956,
71
+ "learning_rate": 8e-05,
72
+ "loss": 1.4883,
73
+ "step": 9
74
+ },
75
+ {
76
+ "epoch": 0.3524229074889868,
77
+ "grad_norm": 0.8478774428367615,
78
+ "learning_rate": 9e-05,
79
+ "loss": 1.3543,
80
+ "step": 10
81
+ },
82
+ {
83
+ "epoch": 0.3524229074889868,
84
+ "eval_loss": 1.2461069822311401,
85
+ "eval_runtime": 25.3282,
86
+ "eval_samples_per_second": 7.975,
87
+ "eval_steps_per_second": 0.513,
88
+ "step": 10
89
+ },
90
+ {
91
+ "epoch": 0.3876651982378855,
92
+ "grad_norm": 0.8932197093963623,
93
+ "learning_rate": 0.0001,
94
+ "loss": 1.253,
95
+ "step": 11
96
+ },
97
+ {
98
+ "epoch": 0.42290748898678415,
99
+ "grad_norm": 0.9408664703369141,
100
+ "learning_rate": 0.00011000000000000002,
101
+ "loss": 1.1504,
102
+ "step": 12
103
+ },
104
+ {
105
+ "epoch": 0.4581497797356828,
106
+ "grad_norm": 0.7685560584068298,
107
+ "learning_rate": 0.00012,
108
+ "loss": 1.1904,
109
+ "step": 13
110
+ },
111
+ {
112
+ "epoch": 0.4933920704845815,
113
+ "grad_norm": 1.0110777616500854,
114
+ "learning_rate": 0.00013000000000000002,
115
+ "loss": 1.065,
116
+ "step": 14
117
+ },
118
+ {
119
+ "epoch": 0.5286343612334802,
120
+ "grad_norm": 1.0698988437652588,
121
+ "learning_rate": 0.00014,
122
+ "loss": 1.0323,
123
+ "step": 15
124
+ },
125
+ {
126
+ "epoch": 0.5638766519823789,
127
+ "grad_norm": 0.9592034220695496,
128
+ "learning_rate": 0.00015000000000000001,
129
+ "loss": 0.967,
130
+ "step": 16
131
+ },
132
+ {
133
+ "epoch": 0.5991189427312775,
134
+ "grad_norm": 0.515692412853241,
135
+ "learning_rate": 0.00016,
136
+ "loss": 0.9886,
137
+ "step": 17
138
+ },
139
+ {
140
+ "epoch": 0.6343612334801763,
141
+ "grad_norm": 0.4879838824272156,
142
+ "learning_rate": 0.00017,
143
+ "loss": 0.9052,
144
+ "step": 18
145
+ },
146
+ {
147
+ "epoch": 0.6696035242290749,
148
+ "grad_norm": 0.5619950294494629,
149
+ "learning_rate": 0.00018,
150
+ "loss": 0.8916,
151
+ "step": 19
152
+ },
153
+ {
154
+ "epoch": 0.7048458149779736,
155
+ "grad_norm": 0.5306786298751831,
156
+ "learning_rate": 0.00019,
157
+ "loss": 0.8508,
158
+ "step": 20
159
+ },
160
+ {
161
+ "epoch": 0.7048458149779736,
162
+ "eval_loss": 0.8256819844245911,
163
+ "eval_runtime": 25.3172,
164
+ "eval_samples_per_second": 7.979,
165
+ "eval_steps_per_second": 0.513,
166
+ "step": 20
167
+ },
168
+ {
169
+ "epoch": 0.7400881057268722,
170
+ "grad_norm": 0.45645713806152344,
171
+ "learning_rate": 0.0002,
172
+ "loss": 0.837,
173
+ "step": 21
174
+ },
175
+ {
176
+ "epoch": 0.775330396475771,
177
+ "grad_norm": 0.3715068995952606,
178
+ "learning_rate": 0.0001999229036240723,
179
+ "loss": 0.8872,
180
+ "step": 22
181
+ },
182
+ {
183
+ "epoch": 0.8105726872246696,
184
+ "grad_norm": 0.3798660337924957,
185
+ "learning_rate": 0.0001996917333733128,
186
+ "loss": 0.8076,
187
+ "step": 23
188
+ },
189
+ {
190
+ "epoch": 0.8458149779735683,
191
+ "grad_norm": 0.4133865237236023,
192
+ "learning_rate": 0.00019930684569549264,
193
+ "loss": 0.8181,
194
+ "step": 24
195
+ },
196
+ {
197
+ "epoch": 0.8810572687224669,
198
+ "grad_norm": 0.3884609043598175,
199
+ "learning_rate": 0.00019876883405951377,
200
+ "loss": 0.8081,
201
+ "step": 25
202
+ },
203
+ {
204
+ "epoch": 0.9162995594713657,
205
+ "grad_norm": 0.3344547748565674,
206
+ "learning_rate": 0.00019807852804032305,
207
+ "loss": 0.7206,
208
+ "step": 26
209
+ },
210
+ {
211
+ "epoch": 0.9515418502202643,
212
+ "grad_norm": 0.2715914845466614,
213
+ "learning_rate": 0.00019723699203976766,
214
+ "loss": 0.7821,
215
+ "step": 27
216
+ },
217
+ {
218
+ "epoch": 0.986784140969163,
219
+ "grad_norm": 0.3191257417201996,
220
+ "learning_rate": 0.00019624552364536473,
221
+ "loss": 0.7356,
222
+ "step": 28
223
+ },
224
+ {
225
+ "epoch": 1.0,
226
+ "grad_norm": 0.43808606266975403,
227
+ "learning_rate": 0.00019510565162951537,
228
+ "loss": 0.7774,
229
+ "step": 29
230
+ },
231
+ {
232
+ "epoch": 1.0352422907488987,
233
+ "grad_norm": 0.32706010341644287,
234
+ "learning_rate": 0.00019381913359224842,
235
+ "loss": 0.7069,
236
+ "step": 30
237
+ },
238
+ {
239
+ "epoch": 1.0352422907488987,
240
+ "eval_loss": 0.6948859691619873,
241
+ "eval_runtime": 25.3163,
242
+ "eval_samples_per_second": 7.979,
243
+ "eval_steps_per_second": 0.514,
244
+ "step": 30
245
+ },
246
+ {
247
+ "epoch": 1.0704845814977975,
248
+ "grad_norm": 0.2936725616455078,
249
+ "learning_rate": 0.0001923879532511287,
250
+ "loss": 0.7228,
251
+ "step": 31
252
+ },
253
+ {
254
+ "epoch": 1.105726872246696,
255
+ "grad_norm": 0.3039261996746063,
256
+ "learning_rate": 0.00019081431738250814,
257
+ "loss": 0.6692,
258
+ "step": 32
259
+ },
260
+ {
261
+ "epoch": 1.1409691629955947,
262
+ "grad_norm": 0.29901978373527527,
263
+ "learning_rate": 0.0001891006524188368,
264
+ "loss": 0.6742,
265
+ "step": 33
266
+ },
267
+ {
268
+ "epoch": 1.1762114537444934,
269
+ "grad_norm": 0.32121536135673523,
270
+ "learning_rate": 0.00018724960070727972,
271
+ "loss": 0.7012,
272
+ "step": 34
273
+ },
274
+ {
275
+ "epoch": 1.2114537444933922,
276
+ "grad_norm": 0.3143290579319,
277
+ "learning_rate": 0.00018526401643540922,
278
+ "loss": 0.6594,
279
+ "step": 35
280
+ },
281
+ {
282
+ "epoch": 1.2466960352422907,
283
+ "grad_norm": 0.3040238320827484,
284
+ "learning_rate": 0.00018314696123025454,
285
+ "loss": 0.6723,
286
+ "step": 36
287
+ },
288
+ {
289
+ "epoch": 1.2819383259911894,
290
+ "grad_norm": 0.2974918782711029,
291
+ "learning_rate": 0.00018090169943749476,
292
+ "loss": 0.6514,
293
+ "step": 37
294
+ },
295
+ {
296
+ "epoch": 1.3171806167400881,
297
+ "grad_norm": 0.2963806390762329,
298
+ "learning_rate": 0.00017853169308807448,
299
+ "loss": 0.6585,
300
+ "step": 38
301
+ },
302
+ {
303
+ "epoch": 1.3524229074889869,
304
+ "grad_norm": 0.2951265871524811,
305
+ "learning_rate": 0.0001760405965600031,
306
+ "loss": 0.6339,
307
+ "step": 39
308
+ },
309
+ {
310
+ "epoch": 1.3876651982378854,
311
+ "grad_norm": 0.3203594386577606,
312
+ "learning_rate": 0.00017343225094356855,
313
+ "loss": 0.655,
314
+ "step": 40
315
+ },
316
+ {
317
+ "epoch": 1.3876651982378854,
318
+ "eval_loss": 0.6199031472206116,
319
+ "eval_runtime": 25.3133,
320
+ "eval_samples_per_second": 7.98,
321
+ "eval_steps_per_second": 0.514,
322
+ "step": 40
323
+ },
324
+ {
325
+ "epoch": 1.422907488986784,
326
+ "grad_norm": 0.29003453254699707,
327
+ "learning_rate": 0.00017071067811865476,
328
+ "loss": 0.6334,
329
+ "step": 41
330
+ },
331
+ {
332
+ "epoch": 1.4581497797356828,
333
+ "grad_norm": 0.29403063654899597,
334
+ "learning_rate": 0.0001678800745532942,
335
+ "loss": 0.644,
336
+ "step": 42
337
+ },
338
+ {
339
+ "epoch": 1.4933920704845816,
340
+ "grad_norm": 0.30763620138168335,
341
+ "learning_rate": 0.00016494480483301836,
342
+ "loss": 0.5925,
343
+ "step": 43
344
+ },
345
+ {
346
+ "epoch": 1.5286343612334803,
347
+ "grad_norm": 0.3372487425804138,
348
+ "learning_rate": 0.00016190939493098344,
349
+ "loss": 0.6486,
350
+ "step": 44
351
+ },
352
+ {
353
+ "epoch": 1.5638766519823788,
354
+ "grad_norm": 0.2898256480693817,
355
+ "learning_rate": 0.00015877852522924732,
356
+ "loss": 0.6039,
357
+ "step": 45
358
+ },
359
+ {
360
+ "epoch": 1.5991189427312775,
361
+ "grad_norm": 0.2907792627811432,
362
+ "learning_rate": 0.00015555702330196023,
363
+ "loss": 0.6244,
364
+ "step": 46
365
+ },
366
+ {
367
+ "epoch": 1.6343612334801763,
368
+ "grad_norm": 0.3023645579814911,
369
+ "learning_rate": 0.0001522498564715949,
370
+ "loss": 0.5672,
371
+ "step": 47
372
+ },
373
+ {
374
+ "epoch": 1.6696035242290748,
375
+ "grad_norm": 0.3130245804786682,
376
+ "learning_rate": 0.00014886212414969553,
377
+ "loss": 0.6356,
378
+ "step": 48
379
+ },
380
+ {
381
+ "epoch": 1.7048458149779737,
382
+ "grad_norm": 0.3192823529243469,
383
+ "learning_rate": 0.00014539904997395468,
384
+ "loss": 0.6105,
385
+ "step": 49
386
+ },
387
+ {
388
+ "epoch": 1.7400881057268722,
389
+ "grad_norm": 0.3187403082847595,
390
+ "learning_rate": 0.0001418659737537428,
391
+ "loss": 0.6023,
392
+ "step": 50
393
+ },
394
+ {
395
+ "epoch": 1.7400881057268722,
396
+ "eval_loss": 0.5753047466278076,
397
+ "eval_runtime": 25.3636,
398
+ "eval_samples_per_second": 7.964,
399
+ "eval_steps_per_second": 0.513,
400
+ "step": 50
401
+ },
402
+ {
403
+ "epoch": 1.775330396475771,
404
+ "grad_norm": 0.3186057507991791,
405
+ "learning_rate": 0.000138268343236509,
406
+ "loss": 0.6216,
407
+ "step": 51
408
+ },
409
+ {
410
+ "epoch": 1.8105726872246697,
411
+ "grad_norm": 0.3059786856174469,
412
+ "learning_rate": 0.0001346117057077493,
413
+ "loss": 0.6124,
414
+ "step": 52
415
+ },
416
+ {
417
+ "epoch": 1.8458149779735682,
418
+ "grad_norm": 0.2928261160850525,
419
+ "learning_rate": 0.00013090169943749476,
420
+ "loss": 0.5445,
421
+ "step": 53
422
+ },
423
+ {
424
+ "epoch": 1.881057268722467,
425
+ "grad_norm": 0.2955411374568939,
426
+ "learning_rate": 0.00012714404498650743,
427
+ "loss": 0.5508,
428
+ "step": 54
429
+ },
430
+ {
431
+ "epoch": 1.9162995594713657,
432
+ "grad_norm": 0.30531296133995056,
433
+ "learning_rate": 0.00012334453638559057,
434
+ "loss": 0.5522,
435
+ "step": 55
436
+ },
437
+ {
438
+ "epoch": 1.9515418502202642,
439
+ "grad_norm": 0.3257448375225067,
440
+ "learning_rate": 0.00011950903220161285,
441
+ "loss": 0.5569,
442
+ "step": 56
443
+ },
444
+ {
445
+ "epoch": 1.9867841409691631,
446
+ "grad_norm": 0.29654088616371155,
447
+ "learning_rate": 0.0001156434465040231,
448
+ "loss": 0.5307,
449
+ "step": 57
450
+ },
451
+ {
452
+ "epoch": 2.0,
453
+ "grad_norm": 0.46664145588874817,
454
+ "learning_rate": 0.00011175373974578378,
455
+ "loss": 0.6131,
456
+ "step": 58
457
+ },
458
+ {
459
+ "epoch": 2.0352422907488985,
460
+ "grad_norm": 0.3071623742580414,
461
+ "learning_rate": 0.0001078459095727845,
462
+ "loss": 0.4965,
463
+ "step": 59
464
+ },
465
+ {
466
+ "epoch": 2.0704845814977975,
467
+ "grad_norm": 0.3435165286064148,
468
+ "learning_rate": 0.00010392598157590688,
469
+ "loss": 0.5086,
470
+ "step": 60
471
+ },
472
+ {
473
+ "epoch": 2.0704845814977975,
474
+ "eval_loss": 0.5449034571647644,
475
+ "eval_runtime": 25.365,
476
+ "eval_samples_per_second": 7.964,
477
+ "eval_steps_per_second": 0.513,
478
+ "step": 60
479
+ },
480
+ {
481
+ "epoch": 2.105726872246696,
482
+ "grad_norm": 0.3053547739982605,
483
+ "learning_rate": 0.0001,
484
+ "loss": 0.5208,
485
+ "step": 61
486
+ },
487
+ {
488
+ "epoch": 2.140969162995595,
489
+ "grad_norm": 0.3197723627090454,
490
+ "learning_rate": 9.607401842409317e-05,
491
+ "loss": 0.5298,
492
+ "step": 62
493
+ },
494
+ {
495
+ "epoch": 2.1762114537444934,
496
+ "grad_norm": 0.31069523096084595,
497
+ "learning_rate": 9.215409042721552e-05,
498
+ "loss": 0.5025,
499
+ "step": 63
500
+ },
501
+ {
502
+ "epoch": 2.211453744493392,
503
+ "grad_norm": 0.3025587797164917,
504
+ "learning_rate": 8.824626025421626e-05,
505
+ "loss": 0.4856,
506
+ "step": 64
507
+ },
508
+ {
509
+ "epoch": 2.246696035242291,
510
+ "grad_norm": 0.32704848051071167,
511
+ "learning_rate": 8.435655349597689e-05,
512
+ "loss": 0.5118,
513
+ "step": 65
514
+ },
515
+ {
516
+ "epoch": 2.2819383259911894,
517
+ "grad_norm": 0.3421045243740082,
518
+ "learning_rate": 8.049096779838719e-05,
519
+ "loss": 0.5256,
520
+ "step": 66
521
+ },
522
+ {
523
+ "epoch": 2.317180616740088,
524
+ "grad_norm": 0.3493560552597046,
525
+ "learning_rate": 7.66554636144095e-05,
526
+ "loss": 0.5396,
527
+ "step": 67
528
+ },
529
+ {
530
+ "epoch": 2.352422907488987,
531
+ "grad_norm": 0.34267452359199524,
532
+ "learning_rate": 7.285595501349258e-05,
533
+ "loss": 0.5633,
534
+ "step": 68
535
+ },
536
+ {
537
+ "epoch": 2.3876651982378854,
538
+ "grad_norm": 0.34627947211265564,
539
+ "learning_rate": 6.909830056250527e-05,
540
+ "loss": 0.5233,
541
+ "step": 69
542
+ },
543
+ {
544
+ "epoch": 2.4229074889867843,
545
+ "grad_norm": 0.3640250265598297,
546
+ "learning_rate": 6.538829429225069e-05,
547
+ "loss": 0.5638,
548
+ "step": 70
549
+ },
550
+ {
551
+ "epoch": 2.4229074889867843,
552
+ "eval_loss": 0.5274443030357361,
553
+ "eval_runtime": 25.3671,
554
+ "eval_samples_per_second": 7.963,
555
+ "eval_steps_per_second": 0.512,
556
+ "step": 70
557
+ },
558
+ {
559
+ "epoch": 2.458149779735683,
560
+ "grad_norm": 0.3336215913295746,
561
+ "learning_rate": 6.173165676349103e-05,
562
+ "loss": 0.5003,
563
+ "step": 71
564
+ },
565
+ {
566
+ "epoch": 2.4933920704845813,
567
+ "grad_norm": 0.3200637102127075,
568
+ "learning_rate": 5.8134026246257225e-05,
569
+ "loss": 0.4974,
570
+ "step": 72
571
+ },
572
+ {
573
+ "epoch": 2.5286343612334803,
574
+ "grad_norm": 0.3232054114341736,
575
+ "learning_rate": 5.4600950026045326e-05,
576
+ "loss": 0.5325,
577
+ "step": 73
578
+ },
579
+ {
580
+ "epoch": 2.563876651982379,
581
+ "grad_norm": 0.327750027179718,
582
+ "learning_rate": 5.113787585030454e-05,
583
+ "loss": 0.4323,
584
+ "step": 74
585
+ },
586
+ {
587
+ "epoch": 2.5991189427312777,
588
+ "grad_norm": 0.325334757566452,
589
+ "learning_rate": 4.7750143528405126e-05,
590
+ "loss": 0.4905,
591
+ "step": 75
592
+ },
593
+ {
594
+ "epoch": 2.6343612334801763,
595
+ "grad_norm": 0.3405950665473938,
596
+ "learning_rate": 4.444297669803981e-05,
597
+ "loss": 0.4649,
598
+ "step": 76
599
+ },
600
+ {
601
+ "epoch": 2.6696035242290748,
602
+ "grad_norm": 0.35329973697662354,
603
+ "learning_rate": 4.12214747707527e-05,
604
+ "loss": 0.5379,
605
+ "step": 77
606
+ },
607
+ {
608
+ "epoch": 2.7048458149779737,
609
+ "grad_norm": 0.34212884306907654,
610
+ "learning_rate": 3.8090605069016595e-05,
611
+ "loss": 0.4885,
612
+ "step": 78
613
+ },
614
+ {
615
+ "epoch": 2.7400881057268722,
616
+ "grad_norm": 0.33281591534614563,
617
+ "learning_rate": 3.5055195166981645e-05,
618
+ "loss": 0.4923,
619
+ "step": 79
620
+ },
621
+ {
622
+ "epoch": 2.7753303964757707,
623
+ "grad_norm": 0.32896408438682556,
624
+ "learning_rate": 3.211992544670582e-05,
625
+ "loss": 0.4747,
626
+ "step": 80
627
+ },
628
+ {
629
+ "epoch": 2.7753303964757707,
630
+ "eval_loss": 0.5167238712310791,
631
+ "eval_runtime": 25.36,
632
+ "eval_samples_per_second": 7.965,
633
+ "eval_steps_per_second": 0.513,
634
+ "step": 80
635
+ },
636
+ {
637
+ "epoch": 2.8105726872246697,
638
+ "grad_norm": 0.34730497002601624,
639
+ "learning_rate": 2.9289321881345254e-05,
640
+ "loss": 0.4887,
641
+ "step": 81
642
+ },
643
+ {
644
+ "epoch": 2.845814977973568,
645
+ "grad_norm": 0.35023897886276245,
646
+ "learning_rate": 2.6567749056431467e-05,
647
+ "loss": 0.4812,
648
+ "step": 82
649
+ },
650
+ {
651
+ "epoch": 2.8810572687224667,
652
+ "grad_norm": 0.34714555740356445,
653
+ "learning_rate": 2.3959403439996907e-05,
654
+ "loss": 0.4947,
655
+ "step": 83
656
+ },
657
+ {
658
+ "epoch": 2.9162995594713657,
659
+ "grad_norm": 0.3485592007637024,
660
+ "learning_rate": 2.146830691192553e-05,
661
+ "loss": 0.5061,
662
+ "step": 84
663
+ },
664
+ {
665
+ "epoch": 2.951541850220264,
666
+ "grad_norm": 0.36132875084877014,
667
+ "learning_rate": 1.9098300562505266e-05,
668
+ "loss": 0.4976,
669
+ "step": 85
670
+ },
671
+ {
672
+ "epoch": 2.986784140969163,
673
+ "grad_norm": 0.3450870215892792,
674
+ "learning_rate": 1.6853038769745467e-05,
675
+ "loss": 0.459,
676
+ "step": 86
677
+ },
678
+ {
679
+ "epoch": 3.0,
680
+ "grad_norm": 0.5119443535804749,
681
+ "learning_rate": 1.4735983564590783e-05,
682
+ "loss": 0.4275,
683
+ "step": 87
684
+ },
685
+ {
686
+ "epoch": 3.0352422907488985,
687
+ "grad_norm": 0.325406551361084,
688
+ "learning_rate": 1.2750399292720283e-05,
689
+ "loss": 0.4549,
690
+ "step": 88
691
+ },
692
+ {
693
+ "epoch": 3.0704845814977975,
694
+ "grad_norm": 0.33169472217559814,
695
+ "learning_rate": 1.0899347581163221e-05,
696
+ "loss": 0.4918,
697
+ "step": 89
698
+ },
699
+ {
700
+ "epoch": 3.105726872246696,
701
+ "grad_norm": 0.3156157433986664,
702
+ "learning_rate": 9.185682617491863e-06,
703
+ "loss": 0.458,
704
+ "step": 90
705
+ },
706
+ {
707
+ "epoch": 3.105726872246696,
708
+ "eval_loss": 0.5118579268455505,
709
+ "eval_runtime": 25.3571,
710
+ "eval_samples_per_second": 7.966,
711
+ "eval_steps_per_second": 0.513,
712
+ "step": 90
713
+ },
714
+ {
715
+ "epoch": 3.140969162995595,
716
+ "grad_norm": 0.34081166982650757,
717
+ "learning_rate": 7.612046748871327e-06,
718
+ "loss": 0.4861,
719
+ "step": 91
720
+ },
721
+ {
722
+ "epoch": 3.1762114537444934,
723
+ "grad_norm": 0.32250353693962097,
724
+ "learning_rate": 6.180866407751595e-06,
725
+ "loss": 0.4881,
726
+ "step": 92
727
+ },
728
+ {
729
+ "epoch": 3.211453744493392,
730
+ "grad_norm": 0.3292059600353241,
731
+ "learning_rate": 4.8943483704846475e-06,
732
+ "loss": 0.4524,
733
+ "step": 93
734
+ },
735
+ {
736
+ "epoch": 3.246696035242291,
737
+ "grad_norm": 0.3330952525138855,
738
+ "learning_rate": 3.7544763546352834e-06,
739
+ "loss": 0.4678,
740
+ "step": 94
741
+ },
742
+ {
743
+ "epoch": 3.2819383259911894,
744
+ "grad_norm": 0.3274575173854828,
745
+ "learning_rate": 2.7630079602323442e-06,
746
+ "loss": 0.5017,
747
+ "step": 95
748
+ },
749
+ {
750
+ "epoch": 3.317180616740088,
751
+ "grad_norm": 0.3240121006965637,
752
+ "learning_rate": 1.921471959676957e-06,
753
+ "loss": 0.4615,
754
+ "step": 96
755
+ },
756
+ {
757
+ "epoch": 3.352422907488987,
758
+ "grad_norm": 0.34071552753448486,
759
+ "learning_rate": 1.231165940486234e-06,
760
+ "loss": 0.4972,
761
+ "step": 97
762
+ },
763
+ {
764
+ "epoch": 3.3876651982378854,
765
+ "grad_norm": 0.3293885588645935,
766
+ "learning_rate": 6.931543045073708e-07,
767
+ "loss": 0.4188,
768
+ "step": 98
769
+ },
770
+ {
771
+ "epoch": 3.4229074889867843,
772
+ "grad_norm": 0.32528501749038696,
773
+ "learning_rate": 3.0826662668720364e-07,
774
+ "loss": 0.4577,
775
+ "step": 99
776
+ },
777
+ {
778
+ "epoch": 3.458149779735683,
779
+ "grad_norm": 0.35076647996902466,
780
+ "learning_rate": 7.709637592770991e-08,
781
+ "loss": 0.4505,
782
+ "step": 100
783
+ },
784
+ {
785
+ "epoch": 3.458149779735683,
786
+ "eval_loss": 0.5112941265106201,
787
+ "eval_runtime": 25.3907,
788
+ "eval_samples_per_second": 7.956,
789
+ "eval_steps_per_second": 0.512,
790
+ "step": 100
791
+ }
792
+ ],
793
+ "logging_steps": 1,
794
+ "max_steps": 100,
795
+ "num_input_tokens_seen": 0,
796
+ "num_train_epochs": 4,
797
+ "save_steps": 30,
798
+ "stateful_callbacks": {
799
+ "TrainerControl": {
800
+ "args": {
801
+ "should_epoch_stop": false,
802
+ "should_evaluate": false,
803
+ "should_log": false,
804
+ "should_save": true,
805
+ "should_training_stop": true
806
+ },
807
+ "attributes": {}
808
+ }
809
+ },
810
+ "total_flos": 5.237513300119388e+17,
811
+ "train_batch_size": 2,
812
+ "trial_name": null,
813
+ "trial_params": null
814
+ }
TextUI-Cap-7B/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fe86ede0ce9f34f10d4d866c3a42dabb60082d4c755df6bd02e30daa6c7e608
3
+ size 6161
TextUI-Cap-7B/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
TextUI-Func-7B/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: other
4
+ base_model: /root/autodl-fs/Qwen2.5-7B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - lora
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: stage1_cap_func
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # stage1_cap_func
18
+
19
+ This model is a fine-tuned version of [/root/autodl-fs/Qwen2.5-7B-Instruct](https://huggingface.co//root/autodl-fs/Qwen2.5-7B-Instruct) on the textui-caption2k and the textui-function6k datasets.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.4610
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 0.0001
41
+ - train_batch_size: 3
42
+ - eval_batch_size: 4
43
+ - seed: 42
44
+ - distributed_type: multi-GPU
45
+ - num_devices: 4
46
+ - gradient_accumulation_steps: 8
47
+ - total_train_batch_size: 96
48
+ - total_eval_batch_size: 16
49
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
50
+ - lr_scheduler_type: cosine
51
+ - lr_scheduler_warmup_steps: 25
52
+ - training_steps: 250
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss |
57
+ |:-------------:|:------:|:----:|:---------------:|
58
+ | 0.6383 | 0.2436 | 25 | 0.6979 |
59
+ | 0.5722 | 0.4872 | 50 | 0.6059 |
60
+ | 0.5268 | 0.7308 | 75 | 0.5485 |
61
+ | 0.492 | 0.9744 | 100 | 0.5159 |
62
+ | 0.4719 | 1.2241 | 125 | 0.4990 |
63
+ | 0.4031 | 1.4677 | 150 | 0.4809 |
64
+ | 0.3825 | 1.7113 | 175 | 0.4702 |
65
+ | 0.3944 | 1.9549 | 200 | 0.4625 |
66
+ | 0.3411 | 2.2046 | 225 | 0.4616 |
67
+ | 0.3588 | 2.4482 | 250 | 0.4610 |
68
+
69
+
70
+ ### Framework versions
71
+
72
+ - PEFT 0.15.1
73
+ - Transformers 4.51.3
74
+ - Pytorch 2.7.0+cu126
75
+ - Datasets 3.5.0
76
+ - Tokenizers 0.21.1
TextUI-Func-7B/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/root/autodl-fs/Qwen2.5-7B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "v_proj",
28
+ "down_proj",
29
+ "o_proj",
30
+ "k_proj",
31
+ "gate_proj",
32
+ "q_proj",
33
+ "up_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
TextUI-Func-7B/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dffb0a3a2b49157c0b836102e436dc96d8935ee000246e14c2cf22a5fe213fe4
3
+ size 161533192
TextUI-Func-7B/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
TextUI-Func-7B/all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.4482338611449452,
3
+ "eval_loss": 0.4610269367694855,
4
+ "eval_runtime": 116.0895,
5
+ "eval_samples_per_second": 7.382,
6
+ "eval_steps_per_second": 0.465,
7
+ "total_flos": 2.3792355349217935e+18,
8
+ "train_loss": 0.49713707935810086,
9
+ "train_runtime": 9531.3935,
10
+ "train_samples_per_second": 2.518,
11
+ "train_steps_per_second": 0.026
12
+ }
TextUI-Func-7B/eval_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.4482338611449452,
3
+ "eval_loss": 0.4610269367694855,
4
+ "eval_runtime": 116.0895,
5
+ "eval_samples_per_second": 7.382,
6
+ "eval_steps_per_second": 0.465
7
+ }
TextUI-Func-7B/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
TextUI-Func-7B/runs/events.out.tfevents.1749786583.autodl-container-40b74f9912-ad7efc16.23496.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ddb596420cd303025a950e744047483d55dd5a7cbf49dc516de41b638596a51
3
+ size 60772
TextUI-Func-7B/runs/events.out.tfevents.1749796231.autodl-container-40b74f9912-ad7efc16.23496.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0b704a2975b9e6417dac74a3b8f4c2c7f43d1219b8d090c8ae51bebbe9d2fb6
3
+ size 359
TextUI-Func-7B/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
TextUI-Func-7B/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
TextUI-Func-7B/tokenizer_config.json ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "right",
206
+ "split_special_tokens": false,
207
+ "tokenizer_class": "Qwen2Tokenizer",
208
+ "unk_token": null
209
+ }
TextUI-Func-7B/train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.4482338611449452,
3
+ "total_flos": 2.3792355349217935e+18,
4
+ "train_loss": 0.49713707935810086,
5
+ "train_runtime": 9531.3935,
6
+ "train_samples_per_second": 2.518,
7
+ "train_steps_per_second": 0.026
8
+ }
TextUI-Func-7B/trainer_log.jsonl ADDED
@@ -0,0 +1,261 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"current_steps": 1, "total_steps": 250, "loss": 0.9971, "lr": 0.0, "epoch": 0.0097442143727162, "percentage": 0.4, "elapsed_time": "0:00:43", "remaining_time": "3:01:51"}
2
+ {"current_steps": 2, "total_steps": 250, "loss": 0.8357, "lr": 4.000000000000001e-06, "epoch": 0.0194884287454324, "percentage": 0.8, "elapsed_time": "0:01:15", "remaining_time": "2:35:36"}
3
+ {"current_steps": 3, "total_steps": 250, "loss": 1.0968, "lr": 8.000000000000001e-06, "epoch": 0.029232643118148598, "percentage": 1.2, "elapsed_time": "0:01:54", "remaining_time": "2:36:38"}
4
+ {"current_steps": 4, "total_steps": 250, "loss": 0.9436, "lr": 1.2e-05, "epoch": 0.0389768574908648, "percentage": 1.6, "elapsed_time": "0:02:19", "remaining_time": "2:22:57"}
5
+ {"current_steps": 5, "total_steps": 250, "loss": 0.8405, "lr": 1.6000000000000003e-05, "epoch": 0.048721071863580996, "percentage": 2.0, "elapsed_time": "0:02:51", "remaining_time": "2:19:45"}
6
+ {"current_steps": 6, "total_steps": 250, "loss": 0.8719, "lr": 2e-05, "epoch": 0.058465286236297195, "percentage": 2.4, "elapsed_time": "0:03:26", "remaining_time": "2:19:55"}
7
+ {"current_steps": 7, "total_steps": 250, "loss": 0.7487, "lr": 2.4e-05, "epoch": 0.0682095006090134, "percentage": 2.8, "elapsed_time": "0:04:00", "remaining_time": "2:19:08"}
8
+ {"current_steps": 8, "total_steps": 250, "loss": 0.8104, "lr": 2.8000000000000003e-05, "epoch": 0.0779537149817296, "percentage": 3.2, "elapsed_time": "0:04:33", "remaining_time": "2:17:50"}
9
+ {"current_steps": 9, "total_steps": 250, "loss": 0.7356, "lr": 3.2000000000000005e-05, "epoch": 0.0876979293544458, "percentage": 3.6, "elapsed_time": "0:05:02", "remaining_time": "2:14:54"}
10
+ {"current_steps": 10, "total_steps": 250, "loss": 0.7642, "lr": 3.6e-05, "epoch": 0.09744214372716199, "percentage": 4.0, "elapsed_time": "0:05:31", "remaining_time": "2:12:41"}
11
+ {"current_steps": 11, "total_steps": 250, "loss": 0.734, "lr": 4e-05, "epoch": 0.1071863580998782, "percentage": 4.4, "elapsed_time": "0:06:08", "remaining_time": "2:13:36"}
12
+ {"current_steps": 12, "total_steps": 250, "loss": 0.7155, "lr": 4.4000000000000006e-05, "epoch": 0.11693057247259439, "percentage": 4.8, "elapsed_time": "0:06:50", "remaining_time": "2:15:45"}
13
+ {"current_steps": 13, "total_steps": 250, "loss": 0.7207, "lr": 4.8e-05, "epoch": 0.12667478684531058, "percentage": 5.2, "elapsed_time": "0:07:23", "remaining_time": "2:14:54"}
14
+ {"current_steps": 14, "total_steps": 250, "loss": 0.7882, "lr": 5.2000000000000004e-05, "epoch": 0.1364190012180268, "percentage": 5.6, "elapsed_time": "0:07:53", "remaining_time": "2:13:09"}
15
+ {"current_steps": 15, "total_steps": 250, "loss": 0.7188, "lr": 5.6000000000000006e-05, "epoch": 0.146163215590743, "percentage": 6.0, "elapsed_time": "0:08:28", "remaining_time": "2:12:53"}
16
+ {"current_steps": 16, "total_steps": 250, "loss": 0.7002, "lr": 6e-05, "epoch": 0.1559074299634592, "percentage": 6.4, "elapsed_time": "0:09:00", "remaining_time": "2:11:48"}
17
+ {"current_steps": 17, "total_steps": 250, "loss": 0.7174, "lr": 6.400000000000001e-05, "epoch": 0.1656516443361754, "percentage": 6.8, "elapsed_time": "0:09:37", "remaining_time": "2:11:49"}
18
+ {"current_steps": 18, "total_steps": 250, "loss": 0.7085, "lr": 6.800000000000001e-05, "epoch": 0.1753958587088916, "percentage": 7.2, "elapsed_time": "0:10:09", "remaining_time": "2:10:50"}
19
+ {"current_steps": 19, "total_steps": 250, "loss": 0.788, "lr": 7.2e-05, "epoch": 0.1851400730816078, "percentage": 7.6, "elapsed_time": "0:10:46", "remaining_time": "2:11:00"}
20
+ {"current_steps": 20, "total_steps": 250, "loss": 0.6491, "lr": 7.6e-05, "epoch": 0.19488428745432398, "percentage": 8.0, "elapsed_time": "0:11:22", "remaining_time": "2:10:44"}
21
+ {"current_steps": 21, "total_steps": 250, "loss": 0.6749, "lr": 8e-05, "epoch": 0.2046285018270402, "percentage": 8.4, "elapsed_time": "0:11:52", "remaining_time": "2:09:26"}
22
+ {"current_steps": 22, "total_steps": 250, "loss": 0.6431, "lr": 8.4e-05, "epoch": 0.2143727161997564, "percentage": 8.8, "elapsed_time": "0:12:30", "remaining_time": "2:09:40"}
23
+ {"current_steps": 23, "total_steps": 250, "loss": 0.6089, "lr": 8.800000000000001e-05, "epoch": 0.2241169305724726, "percentage": 9.2, "elapsed_time": "0:12:59", "remaining_time": "2:08:12"}
24
+ {"current_steps": 24, "total_steps": 250, "loss": 0.605, "lr": 9.200000000000001e-05, "epoch": 0.23386114494518878, "percentage": 9.6, "elapsed_time": "0:13:34", "remaining_time": "2:07:46"}
25
+ {"current_steps": 25, "total_steps": 250, "loss": 0.6383, "lr": 9.6e-05, "epoch": 0.243605359317905, "percentage": 10.0, "elapsed_time": "0:14:05", "remaining_time": "2:06:49"}
26
+ {"current_steps": 25, "total_steps": 250, "eval_loss": 0.6978875398635864, "epoch": 0.243605359317905, "percentage": 10.0, "elapsed_time": "0:16:01", "remaining_time": "2:24:16"}
27
+ {"current_steps": 26, "total_steps": 250, "loss": 0.6573, "lr": 0.0001, "epoch": 0.25334957369062117, "percentage": 10.4, "elapsed_time": "0:16:32", "remaining_time": "2:22:32"}
28
+ {"current_steps": 27, "total_steps": 250, "loss": 0.6503, "lr": 9.999512620046522e-05, "epoch": 0.2630937880633374, "percentage": 10.8, "elapsed_time": "0:17:07", "remaining_time": "2:21:26"}
29
+ {"current_steps": 28, "total_steps": 250, "loss": 0.6056, "lr": 9.998050575201771e-05, "epoch": 0.2728380024360536, "percentage": 11.2, "elapsed_time": "0:17:50", "remaining_time": "2:21:26"}
30
+ {"current_steps": 29, "total_steps": 250, "loss": 0.6291, "lr": 9.995614150494293e-05, "epoch": 0.28258221680876977, "percentage": 11.6, "elapsed_time": "0:18:40", "remaining_time": "2:22:20"}
31
+ {"current_steps": 30, "total_steps": 250, "loss": 0.5929, "lr": 9.992203820909906e-05, "epoch": 0.292326431181486, "percentage": 12.0, "elapsed_time": "0:19:13", "remaining_time": "2:21:02"}
32
+ {"current_steps": 31, "total_steps": 250, "loss": 0.6166, "lr": 9.987820251299122e-05, "epoch": 0.3020706455542022, "percentage": 12.4, "elapsed_time": "0:19:46", "remaining_time": "2:19:39"}
33
+ {"current_steps": 32, "total_steps": 250, "loss": 0.631, "lr": 9.982464296247522e-05, "epoch": 0.3118148599269184, "percentage": 12.8, "elapsed_time": "0:20:18", "remaining_time": "2:18:21"}
34
+ {"current_steps": 33, "total_steps": 250, "loss": 0.6309, "lr": 9.976136999909156e-05, "epoch": 0.3215590742996346, "percentage": 13.2, "elapsed_time": "0:20:51", "remaining_time": "2:17:10"}
35
+ {"current_steps": 34, "total_steps": 250, "loss": 0.6514, "lr": 9.968839595802982e-05, "epoch": 0.3313032886723508, "percentage": 13.6, "elapsed_time": "0:21:27", "remaining_time": "2:16:19"}
36
+ {"current_steps": 35, "total_steps": 250, "loss": 0.5839, "lr": 9.96057350657239e-05, "epoch": 0.341047503045067, "percentage": 14.0, "elapsed_time": "0:21:56", "remaining_time": "2:14:48"}
37
+ {"current_steps": 36, "total_steps": 250, "loss": 0.5927, "lr": 9.951340343707852e-05, "epoch": 0.3507917174177832, "percentage": 14.4, "elapsed_time": "0:22:24", "remaining_time": "2:13:10"}
38
+ {"current_steps": 37, "total_steps": 250, "loss": 0.6388, "lr": 9.941141907232765e-05, "epoch": 0.36053593179049936, "percentage": 14.8, "elapsed_time": "0:22:58", "remaining_time": "2:12:13"}
39
+ {"current_steps": 38, "total_steps": 250, "loss": 0.6241, "lr": 9.929980185352526e-05, "epoch": 0.3702801461632156, "percentage": 15.2, "elapsed_time": "0:23:30", "remaining_time": "2:11:09"}
40
+ {"current_steps": 39, "total_steps": 250, "loss": 0.5968, "lr": 9.917857354066931e-05, "epoch": 0.3800243605359318, "percentage": 15.6, "elapsed_time": "0:24:12", "remaining_time": "2:10:57"}
41
+ {"current_steps": 40, "total_steps": 250, "loss": 0.6051, "lr": 9.904775776745958e-05, "epoch": 0.38976857490864797, "percentage": 16.0, "elapsed_time": "0:24:53", "remaining_time": "2:10:39"}
42
+ {"current_steps": 41, "total_steps": 250, "loss": 0.6269, "lr": 9.890738003669029e-05, "epoch": 0.3995127892813642, "percentage": 16.4, "elapsed_time": "0:25:21", "remaining_time": "2:09:18"}
43
+ {"current_steps": 42, "total_steps": 250, "loss": 0.6137, "lr": 9.875746771527816e-05, "epoch": 0.4092570036540804, "percentage": 16.8, "elapsed_time": "0:26:02", "remaining_time": "2:08:58"}
44
+ {"current_steps": 43, "total_steps": 250, "loss": 0.5831, "lr": 9.859805002892732e-05, "epoch": 0.4190012180267966, "percentage": 17.2, "elapsed_time": "0:26:43", "remaining_time": "2:08:39"}
45
+ {"current_steps": 44, "total_steps": 250, "loss": 0.5671, "lr": 9.842915805643155e-05, "epoch": 0.4287454323995128, "percentage": 17.6, "elapsed_time": "0:27:13", "remaining_time": "2:07:27"}
46
+ {"current_steps": 45, "total_steps": 250, "loss": 0.5605, "lr": 9.825082472361557e-05, "epoch": 0.438489646772229, "percentage": 18.0, "elapsed_time": "0:27:40", "remaining_time": "2:06:05"}
47
+ {"current_steps": 46, "total_steps": 250, "loss": 0.6192, "lr": 9.806308479691595e-05, "epoch": 0.4482338611449452, "percentage": 18.4, "elapsed_time": "0:28:16", "remaining_time": "2:05:24"}
48
+ {"current_steps": 47, "total_steps": 250, "loss": 0.5856, "lr": 9.786597487660337e-05, "epoch": 0.4579780755176614, "percentage": 18.8, "elapsed_time": "0:28:41", "remaining_time": "2:03:55"}
49
+ {"current_steps": 48, "total_steps": 250, "loss": 0.6433, "lr": 9.765953338964735e-05, "epoch": 0.46772228989037756, "percentage": 19.2, "elapsed_time": "0:29:23", "remaining_time": "2:03:39"}
50
+ {"current_steps": 49, "total_steps": 250, "loss": 0.6637, "lr": 9.744380058222483e-05, "epoch": 0.4774665042630938, "percentage": 19.6, "elapsed_time": "0:29:50", "remaining_time": "2:02:26"}
51
+ {"current_steps": 50, "total_steps": 250, "loss": 0.5722, "lr": 9.721881851187406e-05, "epoch": 0.48721071863581, "percentage": 20.0, "elapsed_time": "0:30:24", "remaining_time": "2:01:39"}
52
+ {"current_steps": 50, "total_steps": 250, "eval_loss": 0.6059486865997314, "epoch": 0.48721071863581, "percentage": 20.0, "elapsed_time": "0:32:21", "remaining_time": "2:09:24"}
53
+ {"current_steps": 51, "total_steps": 250, "loss": 0.6223, "lr": 9.698463103929542e-05, "epoch": 0.49695493300852617, "percentage": 20.4, "elapsed_time": "0:32:51", "remaining_time": "2:08:13"}
54
+ {"current_steps": 52, "total_steps": 250, "loss": 0.572, "lr": 9.674128381980072e-05, "epoch": 0.5066991473812423, "percentage": 20.8, "elapsed_time": "0:33:26", "remaining_time": "2:07:18"}
55
+ {"current_steps": 53, "total_steps": 250, "loss": 0.5225, "lr": 9.648882429441257e-05, "epoch": 0.5164433617539586, "percentage": 21.2, "elapsed_time": "0:34:02", "remaining_time": "2:06:33"}
56
+ {"current_steps": 54, "total_steps": 250, "loss": 0.5592, "lr": 9.622730168061567e-05, "epoch": 0.5261875761266748, "percentage": 21.6, "elapsed_time": "0:34:43", "remaining_time": "2:06:02"}
57
+ {"current_steps": 55, "total_steps": 250, "loss": 0.5857, "lr": 9.595676696276172e-05, "epoch": 0.535931790499391, "percentage": 22.0, "elapsed_time": "0:35:18", "remaining_time": "2:05:12"}
58
+ {"current_steps": 56, "total_steps": 250, "loss": 0.5908, "lr": 9.567727288213005e-05, "epoch": 0.5456760048721072, "percentage": 22.4, "elapsed_time": "0:35:52", "remaining_time": "2:04:18"}
59
+ {"current_steps": 57, "total_steps": 250, "loss": 0.5695, "lr": 9.538887392664544e-05, "epoch": 0.5554202192448234, "percentage": 22.8, "elapsed_time": "0:36:26", "remaining_time": "2:03:24"}
60
+ {"current_steps": 58, "total_steps": 250, "loss": 0.5314, "lr": 9.50916263202557e-05, "epoch": 0.5651644336175395, "percentage": 23.2, "elapsed_time": "0:36:55", "remaining_time": "2:02:14"}
61
+ {"current_steps": 59, "total_steps": 250, "loss": 0.6366, "lr": 9.478558801197065e-05, "epoch": 0.5749086479902558, "percentage": 23.6, "elapsed_time": "0:37:27", "remaining_time": "2:01:14"}
62
+ {"current_steps": 60, "total_steps": 250, "loss": 0.4933, "lr": 9.447081866456489e-05, "epoch": 0.584652862362972, "percentage": 24.0, "elapsed_time": "0:37:54", "remaining_time": "2:00:03"}
63
+ {"current_steps": 61, "total_steps": 250, "loss": 0.6337, "lr": 9.414737964294636e-05, "epoch": 0.5943970767356882, "percentage": 24.4, "elapsed_time": "0:38:29", "remaining_time": "1:59:15"}
64
+ {"current_steps": 62, "total_steps": 250, "loss": 0.5714, "lr": 9.381533400219318e-05, "epoch": 0.6041412911084044, "percentage": 24.8, "elapsed_time": "0:39:04", "remaining_time": "1:58:29"}
65
+ {"current_steps": 63, "total_steps": 250, "loss": 0.5756, "lr": 9.347474647526095e-05, "epoch": 0.6138855054811205, "percentage": 25.2, "elapsed_time": "0:39:32", "remaining_time": "1:57:21"}
66
+ {"current_steps": 64, "total_steps": 250, "loss": 0.5635, "lr": 9.312568346036288e-05, "epoch": 0.6236297198538368, "percentage": 25.6, "elapsed_time": "0:39:58", "remaining_time": "1:56:09"}
67
+ {"current_steps": 65, "total_steps": 250, "loss": 0.5508, "lr": 9.276821300802534e-05, "epoch": 0.633373934226553, "percentage": 26.0, "elapsed_time": "0:40:26", "remaining_time": "1:55:05"}
68
+ {"current_steps": 66, "total_steps": 250, "loss": 0.5118, "lr": 9.24024048078213e-05, "epoch": 0.6431181485992692, "percentage": 26.4, "elapsed_time": "0:40:57", "remaining_time": "1:54:10"}
69
+ {"current_steps": 67, "total_steps": 250, "loss": 0.5555, "lr": 9.202833017478422e-05, "epoch": 0.6528623629719854, "percentage": 26.8, "elapsed_time": "0:41:29", "remaining_time": "1:53:18"}
70
+ {"current_steps": 68, "total_steps": 250, "loss": 0.5712, "lr": 9.164606203550497e-05, "epoch": 0.6626065773447016, "percentage": 27.2, "elapsed_time": "0:41:58", "remaining_time": "1:52:19"}
71
+ {"current_steps": 69, "total_steps": 250, "loss": 0.54, "lr": 9.125567491391476e-05, "epoch": 0.6723507917174177, "percentage": 27.6, "elapsed_time": "0:42:34", "remaining_time": "1:51:39"}
72
+ {"current_steps": 70, "total_steps": 250, "loss": 0.5334, "lr": 9.085724491675642e-05, "epoch": 0.682095006090134, "percentage": 28.0, "elapsed_time": "0:42:58", "remaining_time": "1:50:30"}
73
+ {"current_steps": 71, "total_steps": 250, "loss": 0.558, "lr": 9.045084971874738e-05, "epoch": 0.6918392204628502, "percentage": 28.4, "elapsed_time": "0:43:37", "remaining_time": "1:49:59"}
74
+ {"current_steps": 72, "total_steps": 250, "loss": 0.5592, "lr": 9.003656854743667e-05, "epoch": 0.7015834348355664, "percentage": 28.8, "elapsed_time": "0:44:10", "remaining_time": "1:49:12"}
75
+ {"current_steps": 73, "total_steps": 250, "loss": 0.562, "lr": 8.961448216775954e-05, "epoch": 0.7113276492082826, "percentage": 29.2, "elapsed_time": "0:44:36", "remaining_time": "1:48:09"}
76
+ {"current_steps": 74, "total_steps": 250, "loss": 0.5502, "lr": 8.9184672866292e-05, "epoch": 0.7210718635809987, "percentage": 29.6, "elapsed_time": "0:45:06", "remaining_time": "1:47:17"}
77
+ {"current_steps": 75, "total_steps": 250, "loss": 0.5268, "lr": 8.874722443520899e-05, "epoch": 0.730816077953715, "percentage": 30.0, "elapsed_time": "0:45:45", "remaining_time": "1:46:45"}
78
+ {"current_steps": 75, "total_steps": 250, "eval_loss": 0.5484524369239807, "epoch": 0.730816077953715, "percentage": 30.0, "elapsed_time": "0:47:41", "remaining_time": "1:51:17"}
79
+ {"current_steps": 76, "total_steps": 250, "loss": 0.5651, "lr": 8.83022221559489e-05, "epoch": 0.7405602923264312, "percentage": 30.4, "elapsed_time": "0:48:16", "remaining_time": "1:50:30"}
80
+ {"current_steps": 77, "total_steps": 250, "loss": 0.507, "lr": 8.784975278258783e-05, "epoch": 0.7503045066991474, "percentage": 30.8, "elapsed_time": "0:48:47", "remaining_time": "1:49:37"}
81
+ {"current_steps": 78, "total_steps": 250, "loss": 0.5846, "lr": 8.73899045249266e-05, "epoch": 0.7600487210718636, "percentage": 31.2, "elapsed_time": "0:49:31", "remaining_time": "1:49:13"}
82
+ {"current_steps": 79, "total_steps": 250, "loss": 0.4785, "lr": 8.692276703129421e-05, "epoch": 0.7697929354445798, "percentage": 31.6, "elapsed_time": "0:50:01", "remaining_time": "1:48:17"}
83
+ {"current_steps": 80, "total_steps": 250, "loss": 0.5115, "lr": 8.644843137107059e-05, "epoch": 0.7795371498172959, "percentage": 32.0, "elapsed_time": "0:50:38", "remaining_time": "1:47:35"}
84
+ {"current_steps": 81, "total_steps": 250, "loss": 0.4972, "lr": 8.596699001693255e-05, "epoch": 0.7892813641900122, "percentage": 32.4, "elapsed_time": "0:51:05", "remaining_time": "1:46:36"}
85
+ {"current_steps": 82, "total_steps": 250, "loss": 0.5548, "lr": 8.547853682682604e-05, "epoch": 0.7990255785627284, "percentage": 32.8, "elapsed_time": "0:51:36", "remaining_time": "1:45:44"}
86
+ {"current_steps": 83, "total_steps": 250, "loss": 0.5017, "lr": 8.498316702566828e-05, "epoch": 0.8087697929354446, "percentage": 33.2, "elapsed_time": "0:52:08", "remaining_time": "1:44:54"}
87
+ {"current_steps": 84, "total_steps": 250, "loss": 0.5304, "lr": 8.44809771867835e-05, "epoch": 0.8185140073081608, "percentage": 33.6, "elapsed_time": "0:52:39", "remaining_time": "1:44:04"}
88
+ {"current_steps": 85, "total_steps": 250, "loss": 0.4957, "lr": 8.397206521307584e-05, "epoch": 0.8282582216808769, "percentage": 34.0, "elapsed_time": "0:53:28", "remaining_time": "1:43:49"}
89
+ {"current_steps": 86, "total_steps": 250, "loss": 0.4853, "lr": 8.345653031794292e-05, "epoch": 0.8380024360535931, "percentage": 34.4, "elapsed_time": "0:53:58", "remaining_time": "1:42:56"}
90
+ {"current_steps": 87, "total_steps": 250, "loss": 0.499, "lr": 8.293447300593402e-05, "epoch": 0.8477466504263094, "percentage": 34.8, "elapsed_time": "0:54:35", "remaining_time": "1:42:16"}
91
+ {"current_steps": 88, "total_steps": 250, "loss": 0.5061, "lr": 8.240599505315655e-05, "epoch": 0.8574908647990256, "percentage": 35.2, "elapsed_time": "0:55:08", "remaining_time": "1:41:30"}
92
+ {"current_steps": 89, "total_steps": 250, "loss": 0.5139, "lr": 8.18711994874345e-05, "epoch": 0.8672350791717418, "percentage": 35.6, "elapsed_time": "0:55:38", "remaining_time": "1:40:39"}
93
+ {"current_steps": 90, "total_steps": 250, "loss": 0.5355, "lr": 8.133019056822304e-05, "epoch": 0.876979293544458, "percentage": 36.0, "elapsed_time": "0:56:04", "remaining_time": "1:39:40"}
94
+ {"current_steps": 91, "total_steps": 250, "loss": 0.4889, "lr": 8.07830737662829e-05, "epoch": 0.8867235079171741, "percentage": 36.4, "elapsed_time": "0:56:28", "remaining_time": "1:38:40"}
95
+ {"current_steps": 92, "total_steps": 250, "loss": 0.4788, "lr": 8.022995574311876e-05, "epoch": 0.8964677222898904, "percentage": 36.8, "elapsed_time": "0:56:56", "remaining_time": "1:37:47"}
96
+ {"current_steps": 93, "total_steps": 250, "loss": 0.5104, "lr": 7.967094433018508e-05, "epoch": 0.9062119366626066, "percentage": 37.2, "elapsed_time": "0:57:29", "remaining_time": "1:37:03"}
97
+ {"current_steps": 94, "total_steps": 250, "loss": 0.4903, "lr": 7.910614850786448e-05, "epoch": 0.9159561510353228, "percentage": 37.6, "elapsed_time": "0:58:06", "remaining_time": "1:36:26"}
98
+ {"current_steps": 95, "total_steps": 250, "loss": 0.4926, "lr": 7.85356783842216e-05, "epoch": 0.925700365408039, "percentage": 38.0, "elapsed_time": "0:58:35", "remaining_time": "1:35:35"}
99
+ {"current_steps": 96, "total_steps": 250, "loss": 0.5514, "lr": 7.795964517353735e-05, "epoch": 0.9354445797807551, "percentage": 38.4, "elapsed_time": "0:59:09", "remaining_time": "1:34:54"}
100
+ {"current_steps": 97, "total_steps": 250, "loss": 0.4886, "lr": 7.737816117462752e-05, "epoch": 0.9451887941534713, "percentage": 38.8, "elapsed_time": "0:59:54", "remaining_time": "1:34:30"}
101
+ {"current_steps": 98, "total_steps": 250, "loss": 0.4776, "lr": 7.679133974894983e-05, "epoch": 0.9549330085261876, "percentage": 39.2, "elapsed_time": "1:00:20", "remaining_time": "1:33:35"}
102
+ {"current_steps": 99, "total_steps": 250, "loss": 0.5343, "lr": 7.619929529850397e-05, "epoch": 0.9646772228989038, "percentage": 39.6, "elapsed_time": "1:00:55", "remaining_time": "1:32:55"}
103
+ {"current_steps": 100, "total_steps": 250, "loss": 0.492, "lr": 7.560214324352858e-05, "epoch": 0.97442143727162, "percentage": 40.0, "elapsed_time": "1:01:30", "remaining_time": "1:32:15"}
104
+ {"current_steps": 100, "total_steps": 250, "eval_loss": 0.5159465074539185, "epoch": 0.97442143727162, "percentage": 40.0, "elapsed_time": "1:03:26", "remaining_time": "1:35:09"}
105
+ {"current_steps": 101, "total_steps": 250, "loss": 0.5276, "lr": 7.500000000000001e-05, "epoch": 0.9841656516443362, "percentage": 40.4, "elapsed_time": "1:03:53", "remaining_time": "1:34:15"}
106
+ {"current_steps": 102, "total_steps": 250, "loss": 0.4887, "lr": 7.439298295693665e-05, "epoch": 0.9939098660170523, "percentage": 40.8, "elapsed_time": "1:04:23", "remaining_time": "1:33:25"}
107
+ {"current_steps": 103, "total_steps": 250, "loss": 0.896, "lr": 7.378121045351378e-05, "epoch": 1.0097442143727162, "percentage": 41.2, "elapsed_time": "1:05:22", "remaining_time": "1:33:17"}
108
+ {"current_steps": 104, "total_steps": 250, "loss": 0.4999, "lr": 7.316480175599309e-05, "epoch": 1.0194884287454324, "percentage": 41.6, "elapsed_time": "1:05:59", "remaining_time": "1:32:39"}
109
+ {"current_steps": 105, "total_steps": 250, "loss": 0.4633, "lr": 7.254387703447154e-05, "epoch": 1.0292326431181487, "percentage": 42.0, "elapsed_time": "1:06:36", "remaining_time": "1:31:59"}
110
+ {"current_steps": 106, "total_steps": 250, "loss": 0.4491, "lr": 7.191855733945387e-05, "epoch": 1.0389768574908649, "percentage": 42.4, "elapsed_time": "1:07:09", "remaining_time": "1:31:14"}
111
+ {"current_steps": 107, "total_steps": 250, "loss": 0.4508, "lr": 7.128896457825364e-05, "epoch": 1.048721071863581, "percentage": 42.8, "elapsed_time": "1:07:37", "remaining_time": "1:30:22"}
112
+ {"current_steps": 108, "total_steps": 250, "loss": 0.4297, "lr": 7.06552214912271e-05, "epoch": 1.058465286236297, "percentage": 43.2, "elapsed_time": "1:08:08", "remaining_time": "1:29:35"}
113
+ {"current_steps": 109, "total_steps": 250, "loss": 0.4592, "lr": 7.001745162784477e-05, "epoch": 1.0682095006090133, "percentage": 43.6, "elapsed_time": "1:08:34", "remaining_time": "1:28:42"}
114
+ {"current_steps": 110, "total_steps": 250, "loss": 0.4819, "lr": 6.937577932260515e-05, "epoch": 1.0779537149817295, "percentage": 44.0, "elapsed_time": "1:09:11", "remaining_time": "1:28:03"}
115
+ {"current_steps": 111, "total_steps": 250, "loss": 0.4678, "lr": 6.873032967079561e-05, "epoch": 1.0876979293544458, "percentage": 44.4, "elapsed_time": "1:09:43", "remaining_time": "1:27:18"}
116
+ {"current_steps": 112, "total_steps": 250, "loss": 0.4635, "lr": 6.808122850410461e-05, "epoch": 1.097442143727162, "percentage": 44.8, "elapsed_time": "1:10:19", "remaining_time": "1:26:38"}
117
+ {"current_steps": 113, "total_steps": 250, "loss": 0.4307, "lr": 6.742860236609077e-05, "epoch": 1.1071863580998782, "percentage": 45.2, "elapsed_time": "1:10:53", "remaining_time": "1:25:56"}
118
+ {"current_steps": 114, "total_steps": 250, "loss": 0.416, "lr": 6.677257848751277e-05, "epoch": 1.1169305724725944, "percentage": 45.6, "elapsed_time": "1:11:22", "remaining_time": "1:25:08"}
119
+ {"current_steps": 115, "total_steps": 250, "loss": 0.4662, "lr": 6.611328476152557e-05, "epoch": 1.1266747868453106, "percentage": 46.0, "elapsed_time": "1:11:48", "remaining_time": "1:24:17"}
120
+ {"current_steps": 116, "total_steps": 250, "loss": 0.4301, "lr": 6.545084971874738e-05, "epoch": 1.1364190012180269, "percentage": 46.4, "elapsed_time": "1:12:21", "remaining_time": "1:23:34"}
121
+ {"current_steps": 117, "total_steps": 250, "loss": 0.4564, "lr": 6.478540250220234e-05, "epoch": 1.146163215590743, "percentage": 46.8, "elapsed_time": "1:12:51", "remaining_time": "1:22:49"}
122
+ {"current_steps": 118, "total_steps": 250, "loss": 0.502, "lr": 6.411707284214384e-05, "epoch": 1.155907429963459, "percentage": 47.2, "elapsed_time": "1:13:22", "remaining_time": "1:22:04"}
123
+ {"current_steps": 119, "total_steps": 250, "loss": 0.4536, "lr": 6.344599103076329e-05, "epoch": 1.1656516443361755, "percentage": 47.6, "elapsed_time": "1:14:00", "remaining_time": "1:21:28"}
124
+ {"current_steps": 120, "total_steps": 250, "loss": 0.4576, "lr": 6.277228789678953e-05, "epoch": 1.1753958587088915, "percentage": 48.0, "elapsed_time": "1:14:27", "remaining_time": "1:20:40"}
125
+ {"current_steps": 121, "total_steps": 250, "loss": 0.4814, "lr": 6.209609477998338e-05, "epoch": 1.1851400730816077, "percentage": 48.4, "elapsed_time": "1:14:51", "remaining_time": "1:19:48"}
126
+ {"current_steps": 122, "total_steps": 250, "loss": 0.4277, "lr": 6.141754350553279e-05, "epoch": 1.194884287454324, "percentage": 48.8, "elapsed_time": "1:15:28", "remaining_time": "1:19:11"}
127
+ {"current_steps": 123, "total_steps": 250, "loss": 0.5118, "lr": 6.073676635835317e-05, "epoch": 1.2046285018270402, "percentage": 49.2, "elapsed_time": "1:16:08", "remaining_time": "1:18:37"}
128
+ {"current_steps": 124, "total_steps": 250, "loss": 0.4725, "lr": 6.005389605729824e-05, "epoch": 1.2143727161997564, "percentage": 49.6, "elapsed_time": "1:16:45", "remaining_time": "1:18:00"}
129
+ {"current_steps": 125, "total_steps": 250, "loss": 0.4719, "lr": 5.9369065729286245e-05, "epoch": 1.2241169305724726, "percentage": 50.0, "elapsed_time": "1:17:15", "remaining_time": "1:17:15"}
130
+ {"current_steps": 125, "total_steps": 250, "eval_loss": 0.4989548325538635, "epoch": 1.2241169305724726, "percentage": 50.0, "elapsed_time": "1:19:12", "remaining_time": "1:19:12"}
131
+ {"current_steps": 126, "total_steps": 250, "loss": 0.4735, "lr": 5.868240888334653e-05, "epoch": 1.2338611449451888, "percentage": 50.4, "elapsed_time": "1:19:44", "remaining_time": "1:18:28"}
132
+ {"current_steps": 127, "total_steps": 250, "loss": 0.4729, "lr": 5.799405938459175e-05, "epoch": 1.243605359317905, "percentage": 50.8, "elapsed_time": "1:20:20", "remaining_time": "1:17:48"}
133
+ {"current_steps": 128, "total_steps": 250, "loss": 0.4589, "lr": 5.730415142812059e-05, "epoch": 1.2533495736906213, "percentage": 51.2, "elapsed_time": "1:20:56", "remaining_time": "1:17:09"}
134
+ {"current_steps": 129, "total_steps": 250, "loss": 0.4464, "lr": 5.661281951285613e-05, "epoch": 1.2630937880633373, "percentage": 51.6, "elapsed_time": "1:21:22", "remaining_time": "1:16:19"}
135
+ {"current_steps": 130, "total_steps": 250, "loss": 0.4537, "lr": 5.5920198415325064e-05, "epoch": 1.2728380024360537, "percentage": 52.0, "elapsed_time": "1:21:57", "remaining_time": "1:15:38"}
136
+ {"current_steps": 131, "total_steps": 250, "loss": 0.4158, "lr": 5.522642316338268e-05, "epoch": 1.2825822168087697, "percentage": 52.4, "elapsed_time": "1:22:23", "remaining_time": "1:14:50"}
137
+ {"current_steps": 132, "total_steps": 250, "loss": 0.4176, "lr": 5.453162900988902e-05, "epoch": 1.292326431181486, "percentage": 52.8, "elapsed_time": "1:22:50", "remaining_time": "1:14:03"}
138
+ {"current_steps": 133, "total_steps": 250, "loss": 0.4287, "lr": 5.383595140634093e-05, "epoch": 1.3020706455542022, "percentage": 53.2, "elapsed_time": "1:23:16", "remaining_time": "1:13:15"}
139
+ {"current_steps": 134, "total_steps": 250, "loss": 0.4118, "lr": 5.313952597646568e-05, "epoch": 1.3118148599269184, "percentage": 53.6, "elapsed_time": "1:23:47", "remaining_time": "1:12:32"}
140
+ {"current_steps": 135, "total_steps": 250, "loss": 0.3948, "lr": 5.244248848978067e-05, "epoch": 1.3215590742996346, "percentage": 54.0, "elapsed_time": "1:24:26", "remaining_time": "1:11:55"}
141
+ {"current_steps": 136, "total_steps": 250, "loss": 0.4579, "lr": 5.174497483512506e-05, "epoch": 1.3313032886723508, "percentage": 54.4, "elapsed_time": "1:25:06", "remaining_time": "1:11:20"}
142
+ {"current_steps": 137, "total_steps": 250, "loss": 0.4025, "lr": 5.104712099416785e-05, "epoch": 1.341047503045067, "percentage": 54.8, "elapsed_time": "1:25:36", "remaining_time": "1:10:36"}
143
+ {"current_steps": 138, "total_steps": 250, "loss": 0.4956, "lr": 5.034906301489808e-05, "epoch": 1.3507917174177833, "percentage": 55.2, "elapsed_time": "1:26:08", "remaining_time": "1:09:54"}
144
+ {"current_steps": 139, "total_steps": 250, "loss": 0.4336, "lr": 4.965093698510193e-05, "epoch": 1.3605359317904995, "percentage": 55.6, "elapsed_time": "1:26:38", "remaining_time": "1:09:11"}
145
+ {"current_steps": 140, "total_steps": 250, "loss": 0.4456, "lr": 4.895287900583216e-05, "epoch": 1.3702801461632155, "percentage": 56.0, "elapsed_time": "1:27:11", "remaining_time": "1:08:30"}
146
+ {"current_steps": 141, "total_steps": 250, "loss": 0.3996, "lr": 4.825502516487497e-05, "epoch": 1.380024360535932, "percentage": 56.4, "elapsed_time": "1:27:44", "remaining_time": "1:07:49"}
147
+ {"current_steps": 142, "total_steps": 250, "loss": 0.4155, "lr": 4.755751151021934e-05, "epoch": 1.389768574908648, "percentage": 56.8, "elapsed_time": "1:28:15", "remaining_time": "1:07:07"}
148
+ {"current_steps": 143, "total_steps": 250, "loss": 0.4325, "lr": 4.6860474023534335e-05, "epoch": 1.3995127892813641, "percentage": 57.2, "elapsed_time": "1:28:47", "remaining_time": "1:06:25"}
149
+ {"current_steps": 144, "total_steps": 250, "loss": 0.4166, "lr": 4.616404859365907e-05, "epoch": 1.4092570036540804, "percentage": 57.6, "elapsed_time": "1:29:21", "remaining_time": "1:05:46"}
150
+ {"current_steps": 145, "total_steps": 250, "loss": 0.4436, "lr": 4.5468370990111006e-05, "epoch": 1.4190012180267966, "percentage": 58.0, "elapsed_time": "1:29:55", "remaining_time": "1:05:06"}
151
+ {"current_steps": 146, "total_steps": 250, "loss": 0.3795, "lr": 4.477357683661734e-05, "epoch": 1.4287454323995128, "percentage": 58.4, "elapsed_time": "1:30:24", "remaining_time": "1:04:24"}
152
+ {"current_steps": 147, "total_steps": 250, "loss": 0.3896, "lr": 4.407980158467495e-05, "epoch": 1.438489646772229, "percentage": 58.8, "elapsed_time": "1:30:53", "remaining_time": "1:03:40"}
153
+ {"current_steps": 148, "total_steps": 250, "loss": 0.4036, "lr": 4.3387180487143876e-05, "epoch": 1.4482338611449452, "percentage": 59.2, "elapsed_time": "1:31:18", "remaining_time": "1:02:56"}
154
+ {"current_steps": 149, "total_steps": 250, "loss": 0.4451, "lr": 4.269584857187943e-05, "epoch": 1.4579780755176615, "percentage": 59.6, "elapsed_time": "1:31:53", "remaining_time": "1:02:17"}
155
+ {"current_steps": 150, "total_steps": 250, "loss": 0.4031, "lr": 4.2005940615408264e-05, "epoch": 1.4677222898903777, "percentage": 60.0, "elapsed_time": "1:32:20", "remaining_time": "1:01:33"}
156
+ {"current_steps": 150, "total_steps": 250, "eval_loss": 0.4808570146560669, "epoch": 1.4677222898903777, "percentage": 60.0, "elapsed_time": "1:34:16", "remaining_time": "1:02:50"}
157
+ {"current_steps": 151, "total_steps": 250, "loss": 0.462, "lr": 4.131759111665349e-05, "epoch": 1.4774665042630937, "percentage": 60.4, "elapsed_time": "1:34:51", "remaining_time": "1:02:11"}
158
+ {"current_steps": 152, "total_steps": 250, "loss": 0.4265, "lr": 4.063093427071376e-05, "epoch": 1.4872107186358101, "percentage": 60.8, "elapsed_time": "1:35:22", "remaining_time": "1:01:29"}
159
+ {"current_steps": 153, "total_steps": 250, "loss": 0.3964, "lr": 3.9946103942701777e-05, "epoch": 1.4969549330085261, "percentage": 61.2, "elapsed_time": "1:35:59", "remaining_time": "1:00:51"}
160
+ {"current_steps": 154, "total_steps": 250, "loss": 0.4642, "lr": 3.926323364164684e-05, "epoch": 1.5066991473812423, "percentage": 61.6, "elapsed_time": "1:36:27", "remaining_time": "1:00:08"}
161
+ {"current_steps": 155, "total_steps": 250, "loss": 0.4511, "lr": 3.858245649446721e-05, "epoch": 1.5164433617539586, "percentage": 62.0, "elapsed_time": "1:36:57", "remaining_time": "0:59:25"}
162
+ {"current_steps": 156, "total_steps": 250, "loss": 0.388, "lr": 3.790390522001662e-05, "epoch": 1.5261875761266748, "percentage": 62.4, "elapsed_time": "1:37:43", "remaining_time": "0:58:53"}
163
+ {"current_steps": 157, "total_steps": 250, "loss": 0.375, "lr": 3.7227712103210486e-05, "epoch": 1.535931790499391, "percentage": 62.8, "elapsed_time": "1:38:17", "remaining_time": "0:58:13"}
164
+ {"current_steps": 158, "total_steps": 250, "loss": 0.4433, "lr": 3.655400896923672e-05, "epoch": 1.5456760048721072, "percentage": 63.2, "elapsed_time": "1:38:52", "remaining_time": "0:57:34"}
165
+ {"current_steps": 159, "total_steps": 250, "loss": 0.4002, "lr": 3.588292715785617e-05, "epoch": 1.5554202192448234, "percentage": 63.6, "elapsed_time": "1:39:27", "remaining_time": "0:56:55"}
166
+ {"current_steps": 160, "total_steps": 250, "loss": 0.4141, "lr": 3.5214597497797684e-05, "epoch": 1.5651644336175394, "percentage": 64.0, "elapsed_time": "1:40:10", "remaining_time": "0:56:20"}
167
+ {"current_steps": 161, "total_steps": 250, "loss": 0.4277, "lr": 3.4549150281252636e-05, "epoch": 1.5749086479902559, "percentage": 64.4, "elapsed_time": "1:40:44", "remaining_time": "0:55:41"}
168
+ {"current_steps": 162, "total_steps": 250, "loss": 0.3956, "lr": 3.388671523847445e-05, "epoch": 1.5846528623629719, "percentage": 64.8, "elapsed_time": "1:41:16", "remaining_time": "0:55:00"}
169
+ {"current_steps": 163, "total_steps": 250, "loss": 0.4305, "lr": 3.322742151248725e-05, "epoch": 1.5943970767356883, "percentage": 65.2, "elapsed_time": "1:41:48", "remaining_time": "0:54:20"}
170
+ {"current_steps": 164, "total_steps": 250, "loss": 0.4126, "lr": 3.257139763390925e-05, "epoch": 1.6041412911084043, "percentage": 65.6, "elapsed_time": "1:42:25", "remaining_time": "0:53:42"}
171
+ {"current_steps": 165, "total_steps": 250, "loss": 0.404, "lr": 3.1918771495895396e-05, "epoch": 1.6138855054811205, "percentage": 66.0, "elapsed_time": "1:42:53", "remaining_time": "0:53:00"}
172
+ {"current_steps": 166, "total_steps": 250, "loss": 0.4209, "lr": 3.12696703292044e-05, "epoch": 1.6236297198538368, "percentage": 66.4, "elapsed_time": "1:43:28", "remaining_time": "0:52:21"}
173
+ {"current_steps": 167, "total_steps": 250, "loss": 0.4317, "lr": 3.062422067739485e-05, "epoch": 1.633373934226553, "percentage": 66.8, "elapsed_time": "1:44:04", "remaining_time": "0:51:43"}
174
+ {"current_steps": 168, "total_steps": 250, "loss": 0.3964, "lr": 2.9982548372155263e-05, "epoch": 1.6431181485992692, "percentage": 67.2, "elapsed_time": "1:44:37", "remaining_time": "0:51:04"}
175
+ {"current_steps": 169, "total_steps": 250, "loss": 0.3919, "lr": 2.934477850877292e-05, "epoch": 1.6528623629719854, "percentage": 67.6, "elapsed_time": "1:45:06", "remaining_time": "0:50:22"}
176
+ {"current_steps": 170, "total_steps": 250, "loss": 0.4403, "lr": 2.8711035421746367e-05, "epoch": 1.6626065773447016, "percentage": 68.0, "elapsed_time": "1:45:40", "remaining_time": "0:49:43"}
177
+ {"current_steps": 171, "total_steps": 250, "loss": 0.4265, "lr": 2.8081442660546125e-05, "epoch": 1.6723507917174176, "percentage": 68.4, "elapsed_time": "1:46:10", "remaining_time": "0:49:03"}
178
+ {"current_steps": 172, "total_steps": 250, "loss": 0.4612, "lr": 2.7456122965528475e-05, "epoch": 1.682095006090134, "percentage": 68.8, "elapsed_time": "1:46:50", "remaining_time": "0:48:26"}
179
+ {"current_steps": 173, "total_steps": 250, "loss": 0.3983, "lr": 2.6835198244006927e-05, "epoch": 1.69183922046285, "percentage": 69.2, "elapsed_time": "1:47:24", "remaining_time": "0:47:48"}
180
+ {"current_steps": 174, "total_steps": 250, "loss": 0.4338, "lr": 2.6218789546486234e-05, "epoch": 1.7015834348355665, "percentage": 69.6, "elapsed_time": "1:48:00", "remaining_time": "0:47:10"}
181
+ {"current_steps": 175, "total_steps": 250, "loss": 0.3825, "lr": 2.560701704306336e-05, "epoch": 1.7113276492082825, "percentage": 70.0, "elapsed_time": "1:48:40", "remaining_time": "0:46:34"}
182
+ {"current_steps": 175, "total_steps": 250, "eval_loss": 0.47023555636405945, "epoch": 1.7113276492082825, "percentage": 70.0, "elapsed_time": "1:50:36", "remaining_time": "0:47:24"}
183
+ {"current_steps": 176, "total_steps": 250, "loss": 0.4265, "lr": 2.500000000000001e-05, "epoch": 1.7210718635809987, "percentage": 70.4, "elapsed_time": "1:51:21", "remaining_time": "0:46:49"}
184
+ {"current_steps": 177, "total_steps": 250, "loss": 0.4299, "lr": 2.4397856756471432e-05, "epoch": 1.730816077953715, "percentage": 70.8, "elapsed_time": "1:51:56", "remaining_time": "0:46:10"}
185
+ {"current_steps": 178, "total_steps": 250, "loss": 0.4355, "lr": 2.3800704701496053e-05, "epoch": 1.7405602923264312, "percentage": 71.2, "elapsed_time": "1:52:36", "remaining_time": "0:45:32"}
186
+ {"current_steps": 179, "total_steps": 250, "loss": 0.4235, "lr": 2.3208660251050158e-05, "epoch": 1.7503045066991474, "percentage": 71.6, "elapsed_time": "1:53:18", "remaining_time": "0:44:56"}
187
+ {"current_steps": 180, "total_steps": 250, "loss": 0.4078, "lr": 2.2621838825372493e-05, "epoch": 1.7600487210718636, "percentage": 72.0, "elapsed_time": "1:53:51", "remaining_time": "0:44:16"}
188
+ {"current_steps": 181, "total_steps": 250, "loss": 0.4046, "lr": 2.2040354826462668e-05, "epoch": 1.7697929354445798, "percentage": 72.4, "elapsed_time": "1:54:19", "remaining_time": "0:43:34"}
189
+ {"current_steps": 182, "total_steps": 250, "loss": 0.433, "lr": 2.1464321615778422e-05, "epoch": 1.7795371498172958, "percentage": 72.8, "elapsed_time": "1:54:52", "remaining_time": "0:42:55"}
190
+ {"current_steps": 183, "total_steps": 250, "loss": 0.4318, "lr": 2.0893851492135537e-05, "epoch": 1.7892813641900123, "percentage": 73.2, "elapsed_time": "1:55:27", "remaining_time": "0:42:16"}
191
+ {"current_steps": 184, "total_steps": 250, "loss": 0.4184, "lr": 2.0329055669814934e-05, "epoch": 1.7990255785627283, "percentage": 73.6, "elapsed_time": "1:56:04", "remaining_time": "0:41:38"}
192
+ {"current_steps": 185, "total_steps": 250, "loss": 0.4346, "lr": 1.977004425688126e-05, "epoch": 1.8087697929354447, "percentage": 74.0, "elapsed_time": "1:56:31", "remaining_time": "0:40:56"}
193
+ {"current_steps": 186, "total_steps": 250, "loss": 0.4007, "lr": 1.9216926233717085e-05, "epoch": 1.8185140073081607, "percentage": 74.4, "elapsed_time": "1:57:03", "remaining_time": "0:40:16"}
194
+ {"current_steps": 187, "total_steps": 250, "loss": 0.4165, "lr": 1.866980943177699e-05, "epoch": 1.828258221680877, "percentage": 74.8, "elapsed_time": "1:57:28", "remaining_time": "0:39:34"}
195
+ {"current_steps": 188, "total_steps": 250, "loss": 0.378, "lr": 1.8128800512565513e-05, "epoch": 1.8380024360535931, "percentage": 75.2, "elapsed_time": "1:58:01", "remaining_time": "0:38:55"}
196
+ {"current_steps": 189, "total_steps": 250, "loss": 0.3824, "lr": 1.7594004946843456e-05, "epoch": 1.8477466504263094, "percentage": 75.6, "elapsed_time": "1:58:29", "remaining_time": "0:38:14"}
197
+ {"current_steps": 190, "total_steps": 250, "loss": 0.3788, "lr": 1.7065526994065973e-05, "epoch": 1.8574908647990256, "percentage": 76.0, "elapsed_time": "1:58:55", "remaining_time": "0:37:33"}
198
+ {"current_steps": 191, "total_steps": 250, "loss": 0.448, "lr": 1.6543469682057106e-05, "epoch": 1.8672350791717418, "percentage": 76.4, "elapsed_time": "1:59:32", "remaining_time": "0:36:55"}
199
+ {"current_steps": 192, "total_steps": 250, "loss": 0.3816, "lr": 1.602793478692419e-05, "epoch": 1.876979293544458, "percentage": 76.8, "elapsed_time": "1:59:59", "remaining_time": "0:36:14"}
200
+ {"current_steps": 193, "total_steps": 250, "loss": 0.3979, "lr": 1.551902281321651e-05, "epoch": 1.886723507917174, "percentage": 77.2, "elapsed_time": "2:00:34", "remaining_time": "0:35:36"}
201
+ {"current_steps": 194, "total_steps": 250, "loss": 0.3654, "lr": 1.5016832974331724e-05, "epoch": 1.8964677222898905, "percentage": 77.6, "elapsed_time": "2:01:03", "remaining_time": "0:34:56"}
202
+ {"current_steps": 195, "total_steps": 250, "loss": 0.4432, "lr": 1.4521463173173965e-05, "epoch": 1.9062119366626065, "percentage": 78.0, "elapsed_time": "2:01:33", "remaining_time": "0:34:17"}
203
+ {"current_steps": 196, "total_steps": 250, "loss": 0.364, "lr": 1.4033009983067452e-05, "epoch": 1.915956151035323, "percentage": 78.4, "elapsed_time": "2:02:09", "remaining_time": "0:33:39"}
204
+ {"current_steps": 197, "total_steps": 250, "loss": 0.4442, "lr": 1.3551568628929434e-05, "epoch": 1.925700365408039, "percentage": 78.8, "elapsed_time": "2:02:50", "remaining_time": "0:33:03"}
205
+ {"current_steps": 198, "total_steps": 250, "loss": 0.4042, "lr": 1.3077232968705805e-05, "epoch": 1.9354445797807551, "percentage": 79.2, "elapsed_time": "2:03:23", "remaining_time": "0:32:24"}
206
+ {"current_steps": 199, "total_steps": 250, "loss": 0.4252, "lr": 1.2610095475073414e-05, "epoch": 1.9451887941534713, "percentage": 79.6, "elapsed_time": "2:03:56", "remaining_time": "0:31:45"}
207
+ {"current_steps": 200, "total_steps": 250, "loss": 0.3944, "lr": 1.2150247217412186e-05, "epoch": 1.9549330085261876, "percentage": 80.0, "elapsed_time": "2:04:30", "remaining_time": "0:31:07"}
208
+ {"current_steps": 200, "total_steps": 250, "eval_loss": 0.4625219404697418, "epoch": 1.9549330085261876, "percentage": 80.0, "elapsed_time": "2:06:26", "remaining_time": "0:31:36"}
209
+ {"current_steps": 201, "total_steps": 250, "loss": 0.397, "lr": 1.1697777844051105e-05, "epoch": 1.9646772228989038, "percentage": 80.4, "elapsed_time": "2:06:54", "remaining_time": "0:30:56"}
210
+ {"current_steps": 202, "total_steps": 250, "loss": 0.4191, "lr": 1.1252775564791024e-05, "epoch": 1.97442143727162, "percentage": 80.8, "elapsed_time": "2:07:27", "remaining_time": "0:30:17"}
211
+ {"current_steps": 203, "total_steps": 250, "loss": 0.4126, "lr": 1.0815327133708015e-05, "epoch": 1.9841656516443362, "percentage": 81.2, "elapsed_time": "2:08:15", "remaining_time": "0:29:41"}
212
+ {"current_steps": 204, "total_steps": 250, "loss": 0.3936, "lr": 1.0385517832240471e-05, "epoch": 1.9939098660170522, "percentage": 81.6, "elapsed_time": "2:08:50", "remaining_time": "0:29:03"}
213
+ {"current_steps": 205, "total_steps": 250, "loss": 0.8769, "lr": 9.963431452563332e-06, "epoch": 2.009744214372716, "percentage": 82.0, "elapsed_time": "2:09:47", "remaining_time": "0:28:29"}
214
+ {"current_steps": 206, "total_steps": 250, "loss": 0.3756, "lr": 9.549150281252633e-06, "epoch": 2.0194884287454324, "percentage": 82.4, "elapsed_time": "2:10:19", "remaining_time": "0:27:50"}
215
+ {"current_steps": 207, "total_steps": 250, "loss": 0.4197, "lr": 9.142755083243576e-06, "epoch": 2.0292326431181484, "percentage": 82.8, "elapsed_time": "2:10:46", "remaining_time": "0:27:09"}
216
+ {"current_steps": 208, "total_steps": 250, "loss": 0.4043, "lr": 8.744325086085248e-06, "epoch": 2.038976857490865, "percentage": 83.2, "elapsed_time": "2:11:11", "remaining_time": "0:26:29"}
217
+ {"current_steps": 209, "total_steps": 250, "loss": 0.3806, "lr": 8.353937964495029e-06, "epoch": 2.048721071863581, "percentage": 83.6, "elapsed_time": "2:11:56", "remaining_time": "0:25:53"}
218
+ {"current_steps": 210, "total_steps": 250, "loss": 0.3838, "lr": 7.971669825215788e-06, "epoch": 2.0584652862362973, "percentage": 84.0, "elapsed_time": "2:12:22", "remaining_time": "0:25:12"}
219
+ {"current_steps": 211, "total_steps": 250, "loss": 0.3778, "lr": 7.597595192178702e-06, "epoch": 2.0682095006090133, "percentage": 84.4, "elapsed_time": "2:12:50", "remaining_time": "0:24:33"}
220
+ {"current_steps": 212, "total_steps": 250, "loss": 0.4119, "lr": 7.2317869919746705e-06, "epoch": 2.0779537149817298, "percentage": 84.8, "elapsed_time": "2:13:30", "remaining_time": "0:23:55"}
221
+ {"current_steps": 213, "total_steps": 250, "loss": 0.3804, "lr": 6.874316539637127e-06, "epoch": 2.0876979293544458, "percentage": 85.2, "elapsed_time": "2:14:04", "remaining_time": "0:23:17"}
222
+ {"current_steps": 214, "total_steps": 250, "loss": 0.3932, "lr": 6.52525352473905e-06, "epoch": 2.097442143727162, "percentage": 85.6, "elapsed_time": "2:14:38", "remaining_time": "0:22:39"}
223
+ {"current_steps": 215, "total_steps": 250, "loss": 0.3658, "lr": 6.184665997806832e-06, "epoch": 2.107186358099878, "percentage": 86.0, "elapsed_time": "2:15:04", "remaining_time": "0:21:59"}
224
+ {"current_steps": 216, "total_steps": 250, "loss": 0.3683, "lr": 5.852620357053651e-06, "epoch": 2.116930572472594, "percentage": 86.4, "elapsed_time": "2:15:40", "remaining_time": "0:21:21"}
225
+ {"current_steps": 217, "total_steps": 250, "loss": 0.3635, "lr": 5.529181335435124e-06, "epoch": 2.1266747868453106, "percentage": 86.8, "elapsed_time": "2:16:14", "remaining_time": "0:20:43"}
226
+ {"current_steps": 218, "total_steps": 250, "loss": 0.3929, "lr": 5.214411988029355e-06, "epoch": 2.1364190012180266, "percentage": 87.2, "elapsed_time": "2:16:59", "remaining_time": "0:20:06"}
227
+ {"current_steps": 219, "total_steps": 250, "loss": 0.3579, "lr": 4.908373679744316e-06, "epoch": 2.146163215590743, "percentage": 87.6, "elapsed_time": "2:17:36", "remaining_time": "0:19:28"}
228
+ {"current_steps": 220, "total_steps": 250, "loss": 0.397, "lr": 4.611126073354571e-06, "epoch": 2.155907429963459, "percentage": 88.0, "elapsed_time": "2:18:07", "remaining_time": "0:18:50"}
229
+ {"current_steps": 221, "total_steps": 250, "loss": 0.3996, "lr": 4.322727117869951e-06, "epoch": 2.1656516443361755, "percentage": 88.4, "elapsed_time": "2:18:44", "remaining_time": "0:18:12"}
230
+ {"current_steps": 222, "total_steps": 250, "loss": 0.3797, "lr": 4.043233037238281e-06, "epoch": 2.1753958587088915, "percentage": 88.8, "elapsed_time": "2:19:25", "remaining_time": "0:17:35"}
231
+ {"current_steps": 223, "total_steps": 250, "loss": 0.4177, "lr": 3.772698319384349e-06, "epoch": 2.185140073081608, "percentage": 89.2, "elapsed_time": "2:19:57", "remaining_time": "0:16:56"}
232
+ {"current_steps": 224, "total_steps": 250, "loss": 0.4057, "lr": 3.511175705587433e-06, "epoch": 2.194884287454324, "percentage": 89.6, "elapsed_time": "2:20:38", "remaining_time": "0:16:19"}
233
+ {"current_steps": 225, "total_steps": 250, "loss": 0.3411, "lr": 3.258716180199278e-06, "epoch": 2.2046285018270404, "percentage": 90.0, "elapsed_time": "2:21:11", "remaining_time": "0:15:41"}
234
+ {"current_steps": 225, "total_steps": 250, "eval_loss": 0.461640328168869, "epoch": 2.2046285018270404, "percentage": 90.0, "elapsed_time": "2:23:08", "remaining_time": "0:15:54"}
235
+ {"current_steps": 226, "total_steps": 250, "loss": 0.3767, "lr": 3.0153689607045845e-06, "epoch": 2.2143727161997564, "percentage": 90.4, "elapsed_time": "2:23:38", "remaining_time": "0:15:15"}
236
+ {"current_steps": 227, "total_steps": 250, "loss": 0.3909, "lr": 2.7811814881259503e-06, "epoch": 2.2241169305724724, "percentage": 90.8, "elapsed_time": "2:24:06", "remaining_time": "0:14:36"}
237
+ {"current_steps": 228, "total_steps": 250, "loss": 0.377, "lr": 2.5561994177751737e-06, "epoch": 2.233861144945189, "percentage": 91.2, "elapsed_time": "2:24:35", "remaining_time": "0:13:57"}
238
+ {"current_steps": 229, "total_steps": 250, "loss": 0.3621, "lr": 2.340466610352654e-06, "epoch": 2.243605359317905, "percentage": 91.6, "elapsed_time": "2:25:02", "remaining_time": "0:13:18"}
239
+ {"current_steps": 230, "total_steps": 250, "loss": 0.4011, "lr": 2.134025123396638e-06, "epoch": 2.2533495736906213, "percentage": 92.0, "elapsed_time": "2:25:43", "remaining_time": "0:12:40"}
240
+ {"current_steps": 231, "total_steps": 250, "loss": 0.3517, "lr": 1.9369152030840556e-06, "epoch": 2.2630937880633373, "percentage": 92.4, "elapsed_time": "2:26:13", "remaining_time": "0:12:01"}
241
+ {"current_steps": 232, "total_steps": 250, "loss": 0.3749, "lr": 1.7491752763844293e-06, "epoch": 2.2728380024360537, "percentage": 92.8, "elapsed_time": "2:26:56", "remaining_time": "0:11:24"}
242
+ {"current_steps": 233, "total_steps": 250, "loss": 0.3629, "lr": 1.5708419435684462e-06, "epoch": 2.2825822168087697, "percentage": 93.2, "elapsed_time": "2:27:25", "remaining_time": "0:10:45"}
243
+ {"current_steps": 234, "total_steps": 250, "loss": 0.3971, "lr": 1.4019499710726913e-06, "epoch": 2.292326431181486, "percentage": 93.6, "elapsed_time": "2:28:03", "remaining_time": "0:10:07"}
244
+ {"current_steps": 235, "total_steps": 250, "loss": 0.3643, "lr": 1.2425322847218368e-06, "epoch": 2.302070645554202, "percentage": 94.0, "elapsed_time": "2:28:28", "remaining_time": "0:09:28"}
245
+ {"current_steps": 236, "total_steps": 250, "loss": 0.3678, "lr": 1.0926199633097157e-06, "epoch": 2.311814859926918, "percentage": 94.4, "elapsed_time": "2:29:11", "remaining_time": "0:08:51"}
246
+ {"current_steps": 237, "total_steps": 250, "loss": 0.4407, "lr": 9.522422325404235e-07, "epoch": 2.3215590742996346, "percentage": 94.8, "elapsed_time": "2:29:52", "remaining_time": "0:08:13"}
247
+ {"current_steps": 238, "total_steps": 250, "loss": 0.3228, "lr": 8.214264593307098e-07, "epoch": 2.331303288672351, "percentage": 95.2, "elapsed_time": "2:30:18", "remaining_time": "0:07:34"}
248
+ {"current_steps": 239, "total_steps": 250, "loss": 0.3825, "lr": 7.001981464747565e-07, "epoch": 2.341047503045067, "percentage": 95.6, "elapsed_time": "2:30:52", "remaining_time": "0:06:56"}
249
+ {"current_steps": 240, "total_steps": 250, "loss": 0.3306, "lr": 5.885809276723608e-07, "epoch": 2.350791717417783, "percentage": 96.0, "elapsed_time": "2:31:20", "remaining_time": "0:06:18"}
250
+ {"current_steps": 241, "total_steps": 250, "loss": 0.368, "lr": 4.865965629214819e-07, "epoch": 2.3605359317904995, "percentage": 96.4, "elapsed_time": "2:31:50", "remaining_time": "0:05:40"}
251
+ {"current_steps": 242, "total_steps": 250, "loss": 0.3755, "lr": 3.9426493427611177e-07, "epoch": 2.3702801461632155, "percentage": 96.8, "elapsed_time": "2:32:20", "remaining_time": "0:05:02"}
252
+ {"current_steps": 243, "total_steps": 250, "loss": 0.389, "lr": 3.1160404197018154e-07, "epoch": 2.380024360535932, "percentage": 97.2, "elapsed_time": "2:32:59", "remaining_time": "0:04:24"}
253
+ {"current_steps": 244, "total_steps": 250, "loss": 0.3565, "lr": 2.386300009084408e-07, "epoch": 2.389768574908648, "percentage": 97.6, "elapsed_time": "2:33:33", "remaining_time": "0:03:46"}
254
+ {"current_steps": 245, "total_steps": 250, "loss": 0.3858, "lr": 1.753570375247815e-07, "epoch": 2.3995127892813644, "percentage": 98.0, "elapsed_time": "2:34:11", "remaining_time": "0:03:08"}
255
+ {"current_steps": 246, "total_steps": 250, "loss": 0.3403, "lr": 1.2179748700879012e-07, "epoch": 2.4092570036540804, "percentage": 98.4, "elapsed_time": "2:34:37", "remaining_time": "0:02:30"}
256
+ {"current_steps": 247, "total_steps": 250, "loss": 0.3451, "lr": 7.796179090094891e-08, "epoch": 2.4190012180267964, "percentage": 98.8, "elapsed_time": "2:35:10", "remaining_time": "0:01:53"}
257
+ {"current_steps": 248, "total_steps": 250, "loss": 0.3843, "lr": 4.385849505708084e-08, "epoch": 2.428745432399513, "percentage": 99.2, "elapsed_time": "2:35:47", "remaining_time": "0:01:15"}
258
+ {"current_steps": 249, "total_steps": 250, "loss": 0.409, "lr": 1.949424798228239e-08, "epoch": 2.4384896467722292, "percentage": 99.6, "elapsed_time": "2:36:24", "remaining_time": "0:00:37"}
259
+ {"current_steps": 250, "total_steps": 250, "loss": 0.3588, "lr": 4.873799534788059e-09, "epoch": 2.4482338611449452, "percentage": 100.0, "elapsed_time": "2:36:54", "remaining_time": "0:00:00"}
260
+ {"current_steps": 250, "total_steps": 250, "eval_loss": 0.4610269367694855, "epoch": 2.4482338611449452, "percentage": 100.0, "elapsed_time": "2:38:50", "remaining_time": "0:00:00"}
261
+ {"current_steps": 250, "total_steps": 250, "epoch": 2.4482338611449452, "percentage": 100.0, "elapsed_time": "2:38:51", "remaining_time": "0:00:00"}
TextUI-Func-7B/trainer_state.json ADDED
@@ -0,0 +1,1873 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 250,
3
+ "best_metric": 0.4610269367694855,
4
+ "best_model_checkpoint": "/root/autodl-tmp/model/lora-textui/stage1_cap_func/checkpoint-250",
5
+ "epoch": 2.4482338611449452,
6
+ "eval_steps": 25,
7
+ "global_step": 250,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.0097442143727162,
14
+ "grad_norm": 1.4819260835647583,
15
+ "learning_rate": 0.0,
16
+ "loss": 0.9971,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.0194884287454324,
21
+ "grad_norm": 1.1177597045898438,
22
+ "learning_rate": 4.000000000000001e-06,
23
+ "loss": 0.8357,
24
+ "step": 2
25
+ },
26
+ {
27
+ "epoch": 0.029232643118148598,
28
+ "grad_norm": 1.5783437490463257,
29
+ "learning_rate": 8.000000000000001e-06,
30
+ "loss": 1.0968,
31
+ "step": 3
32
+ },
33
+ {
34
+ "epoch": 0.0389768574908648,
35
+ "grad_norm": 1.3322019577026367,
36
+ "learning_rate": 1.2e-05,
37
+ "loss": 0.9436,
38
+ "step": 4
39
+ },
40
+ {
41
+ "epoch": 0.048721071863580996,
42
+ "grad_norm": 1.1600998640060425,
43
+ "learning_rate": 1.6000000000000003e-05,
44
+ "loss": 0.8405,
45
+ "step": 5
46
+ },
47
+ {
48
+ "epoch": 0.058465286236297195,
49
+ "grad_norm": 3.9179673194885254,
50
+ "learning_rate": 2e-05,
51
+ "loss": 0.8719,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 0.0682095006090134,
56
+ "grad_norm": 0.7671970725059509,
57
+ "learning_rate": 2.4e-05,
58
+ "loss": 0.7487,
59
+ "step": 7
60
+ },
61
+ {
62
+ "epoch": 0.0779537149817296,
63
+ "grad_norm": 0.6959292888641357,
64
+ "learning_rate": 2.8000000000000003e-05,
65
+ "loss": 0.8104,
66
+ "step": 8
67
+ },
68
+ {
69
+ "epoch": 0.0876979293544458,
70
+ "grad_norm": 0.4416683316230774,
71
+ "learning_rate": 3.2000000000000005e-05,
72
+ "loss": 0.7356,
73
+ "step": 9
74
+ },
75
+ {
76
+ "epoch": 0.09744214372716199,
77
+ "grad_norm": 0.48265624046325684,
78
+ "learning_rate": 3.6e-05,
79
+ "loss": 0.7642,
80
+ "step": 10
81
+ },
82
+ {
83
+ "epoch": 0.1071863580998782,
84
+ "grad_norm": 0.45329996943473816,
85
+ "learning_rate": 4e-05,
86
+ "loss": 0.734,
87
+ "step": 11
88
+ },
89
+ {
90
+ "epoch": 0.11693057247259439,
91
+ "grad_norm": 0.506803572177887,
92
+ "learning_rate": 4.4000000000000006e-05,
93
+ "loss": 0.7155,
94
+ "step": 12
95
+ },
96
+ {
97
+ "epoch": 0.12667478684531058,
98
+ "grad_norm": 0.5821135640144348,
99
+ "learning_rate": 4.8e-05,
100
+ "loss": 0.7207,
101
+ "step": 13
102
+ },
103
+ {
104
+ "epoch": 0.1364190012180268,
105
+ "grad_norm": 0.6181040406227112,
106
+ "learning_rate": 5.2000000000000004e-05,
107
+ "loss": 0.7882,
108
+ "step": 14
109
+ },
110
+ {
111
+ "epoch": 0.146163215590743,
112
+ "grad_norm": 0.47155848145484924,
113
+ "learning_rate": 5.6000000000000006e-05,
114
+ "loss": 0.7188,
115
+ "step": 15
116
+ },
117
+ {
118
+ "epoch": 0.1559074299634592,
119
+ "grad_norm": 0.3960055112838745,
120
+ "learning_rate": 6e-05,
121
+ "loss": 0.7002,
122
+ "step": 16
123
+ },
124
+ {
125
+ "epoch": 0.1656516443361754,
126
+ "grad_norm": 0.4034963846206665,
127
+ "learning_rate": 6.400000000000001e-05,
128
+ "loss": 0.7174,
129
+ "step": 17
130
+ },
131
+ {
132
+ "epoch": 0.1753958587088916,
133
+ "grad_norm": 0.408160924911499,
134
+ "learning_rate": 6.800000000000001e-05,
135
+ "loss": 0.7085,
136
+ "step": 18
137
+ },
138
+ {
139
+ "epoch": 0.1851400730816078,
140
+ "grad_norm": 0.4035159647464752,
141
+ "learning_rate": 7.2e-05,
142
+ "loss": 0.788,
143
+ "step": 19
144
+ },
145
+ {
146
+ "epoch": 0.19488428745432398,
147
+ "grad_norm": 0.38708168268203735,
148
+ "learning_rate": 7.6e-05,
149
+ "loss": 0.6491,
150
+ "step": 20
151
+ },
152
+ {
153
+ "epoch": 0.2046285018270402,
154
+ "grad_norm": 0.42144861817359924,
155
+ "learning_rate": 8e-05,
156
+ "loss": 0.6749,
157
+ "step": 21
158
+ },
159
+ {
160
+ "epoch": 0.2143727161997564,
161
+ "grad_norm": 0.3642968237400055,
162
+ "learning_rate": 8.4e-05,
163
+ "loss": 0.6431,
164
+ "step": 22
165
+ },
166
+ {
167
+ "epoch": 0.2241169305724726,
168
+ "grad_norm": 0.4132453203201294,
169
+ "learning_rate": 8.800000000000001e-05,
170
+ "loss": 0.6089,
171
+ "step": 23
172
+ },
173
+ {
174
+ "epoch": 0.23386114494518878,
175
+ "grad_norm": 0.4148790240287781,
176
+ "learning_rate": 9.200000000000001e-05,
177
+ "loss": 0.605,
178
+ "step": 24
179
+ },
180
+ {
181
+ "epoch": 0.243605359317905,
182
+ "grad_norm": 0.46465402841567993,
183
+ "learning_rate": 9.6e-05,
184
+ "loss": 0.6383,
185
+ "step": 25
186
+ },
187
+ {
188
+ "epoch": 0.243605359317905,
189
+ "eval_loss": 0.6978875398635864,
190
+ "eval_runtime": 116.3355,
191
+ "eval_samples_per_second": 7.367,
192
+ "eval_steps_per_second": 0.464,
193
+ "step": 25
194
+ },
195
+ {
196
+ "epoch": 0.25334957369062117,
197
+ "grad_norm": 0.38448578119277954,
198
+ "learning_rate": 0.0001,
199
+ "loss": 0.6573,
200
+ "step": 26
201
+ },
202
+ {
203
+ "epoch": 0.2630937880633374,
204
+ "grad_norm": 0.386096715927124,
205
+ "learning_rate": 9.999512620046522e-05,
206
+ "loss": 0.6503,
207
+ "step": 27
208
+ },
209
+ {
210
+ "epoch": 0.2728380024360536,
211
+ "grad_norm": 0.4173428416252136,
212
+ "learning_rate": 9.998050575201771e-05,
213
+ "loss": 0.6056,
214
+ "step": 28
215
+ },
216
+ {
217
+ "epoch": 0.28258221680876977,
218
+ "grad_norm": 0.3838996887207031,
219
+ "learning_rate": 9.995614150494293e-05,
220
+ "loss": 0.6291,
221
+ "step": 29
222
+ },
223
+ {
224
+ "epoch": 0.292326431181486,
225
+ "grad_norm": 0.37374845147132874,
226
+ "learning_rate": 9.992203820909906e-05,
227
+ "loss": 0.5929,
228
+ "step": 30
229
+ },
230
+ {
231
+ "epoch": 0.3020706455542022,
232
+ "grad_norm": 0.38990911841392517,
233
+ "learning_rate": 9.987820251299122e-05,
234
+ "loss": 0.6166,
235
+ "step": 31
236
+ },
237
+ {
238
+ "epoch": 0.3118148599269184,
239
+ "grad_norm": 0.3717946410179138,
240
+ "learning_rate": 9.982464296247522e-05,
241
+ "loss": 0.631,
242
+ "step": 32
243
+ },
244
+ {
245
+ "epoch": 0.3215590742996346,
246
+ "grad_norm": 0.38515135645866394,
247
+ "learning_rate": 9.976136999909156e-05,
248
+ "loss": 0.6309,
249
+ "step": 33
250
+ },
251
+ {
252
+ "epoch": 0.3313032886723508,
253
+ "grad_norm": 0.4617915749549866,
254
+ "learning_rate": 9.968839595802982e-05,
255
+ "loss": 0.6514,
256
+ "step": 34
257
+ },
258
+ {
259
+ "epoch": 0.341047503045067,
260
+ "grad_norm": 0.36573508381843567,
261
+ "learning_rate": 9.96057350657239e-05,
262
+ "loss": 0.5839,
263
+ "step": 35
264
+ },
265
+ {
266
+ "epoch": 0.3507917174177832,
267
+ "grad_norm": 0.38394173979759216,
268
+ "learning_rate": 9.951340343707852e-05,
269
+ "loss": 0.5927,
270
+ "step": 36
271
+ },
272
+ {
273
+ "epoch": 0.36053593179049936,
274
+ "grad_norm": 0.41442376375198364,
275
+ "learning_rate": 9.941141907232765e-05,
276
+ "loss": 0.6388,
277
+ "step": 37
278
+ },
279
+ {
280
+ "epoch": 0.3702801461632156,
281
+ "grad_norm": 0.3883739709854126,
282
+ "learning_rate": 9.929980185352526e-05,
283
+ "loss": 0.6241,
284
+ "step": 38
285
+ },
286
+ {
287
+ "epoch": 0.3800243605359318,
288
+ "grad_norm": 0.3784323036670685,
289
+ "learning_rate": 9.917857354066931e-05,
290
+ "loss": 0.5968,
291
+ "step": 39
292
+ },
293
+ {
294
+ "epoch": 0.38976857490864797,
295
+ "grad_norm": 0.38477954268455505,
296
+ "learning_rate": 9.904775776745958e-05,
297
+ "loss": 0.6051,
298
+ "step": 40
299
+ },
300
+ {
301
+ "epoch": 0.3995127892813642,
302
+ "grad_norm": 0.42043524980545044,
303
+ "learning_rate": 9.890738003669029e-05,
304
+ "loss": 0.6269,
305
+ "step": 41
306
+ },
307
+ {
308
+ "epoch": 0.4092570036540804,
309
+ "grad_norm": 0.4155956208705902,
310
+ "learning_rate": 9.875746771527816e-05,
311
+ "loss": 0.6137,
312
+ "step": 42
313
+ },
314
+ {
315
+ "epoch": 0.4190012180267966,
316
+ "grad_norm": 0.39818403124809265,
317
+ "learning_rate": 9.859805002892732e-05,
318
+ "loss": 0.5831,
319
+ "step": 43
320
+ },
321
+ {
322
+ "epoch": 0.4287454323995128,
323
+ "grad_norm": 0.4027486741542816,
324
+ "learning_rate": 9.842915805643155e-05,
325
+ "loss": 0.5671,
326
+ "step": 44
327
+ },
328
+ {
329
+ "epoch": 0.438489646772229,
330
+ "grad_norm": 0.3745126724243164,
331
+ "learning_rate": 9.825082472361557e-05,
332
+ "loss": 0.5605,
333
+ "step": 45
334
+ },
335
+ {
336
+ "epoch": 0.4482338611449452,
337
+ "grad_norm": 0.4132031798362732,
338
+ "learning_rate": 9.806308479691595e-05,
339
+ "loss": 0.6192,
340
+ "step": 46
341
+ },
342
+ {
343
+ "epoch": 0.4579780755176614,
344
+ "grad_norm": 0.42464327812194824,
345
+ "learning_rate": 9.786597487660337e-05,
346
+ "loss": 0.5856,
347
+ "step": 47
348
+ },
349
+ {
350
+ "epoch": 0.46772228989037756,
351
+ "grad_norm": 0.42480695247650146,
352
+ "learning_rate": 9.765953338964735e-05,
353
+ "loss": 0.6433,
354
+ "step": 48
355
+ },
356
+ {
357
+ "epoch": 0.4774665042630938,
358
+ "grad_norm": 0.4505927562713623,
359
+ "learning_rate": 9.744380058222483e-05,
360
+ "loss": 0.6637,
361
+ "step": 49
362
+ },
363
+ {
364
+ "epoch": 0.48721071863581,
365
+ "grad_norm": 0.40766066312789917,
366
+ "learning_rate": 9.721881851187406e-05,
367
+ "loss": 0.5722,
368
+ "step": 50
369
+ },
370
+ {
371
+ "epoch": 0.48721071863581,
372
+ "eval_loss": 0.6059486865997314,
373
+ "eval_runtime": 116.3481,
374
+ "eval_samples_per_second": 7.366,
375
+ "eval_steps_per_second": 0.464,
376
+ "step": 50
377
+ },
378
+ {
379
+ "epoch": 0.49695493300852617,
380
+ "grad_norm": 0.45385217666625977,
381
+ "learning_rate": 9.698463103929542e-05,
382
+ "loss": 0.6223,
383
+ "step": 51
384
+ },
385
+ {
386
+ "epoch": 0.5066991473812423,
387
+ "grad_norm": 0.4196796417236328,
388
+ "learning_rate": 9.674128381980072e-05,
389
+ "loss": 0.572,
390
+ "step": 52
391
+ },
392
+ {
393
+ "epoch": 0.5164433617539586,
394
+ "grad_norm": 0.4284449517726898,
395
+ "learning_rate": 9.648882429441257e-05,
396
+ "loss": 0.5225,
397
+ "step": 53
398
+ },
399
+ {
400
+ "epoch": 0.5261875761266748,
401
+ "grad_norm": 0.45584678649902344,
402
+ "learning_rate": 9.622730168061567e-05,
403
+ "loss": 0.5592,
404
+ "step": 54
405
+ },
406
+ {
407
+ "epoch": 0.535931790499391,
408
+ "grad_norm": 0.4218933582305908,
409
+ "learning_rate": 9.595676696276172e-05,
410
+ "loss": 0.5857,
411
+ "step": 55
412
+ },
413
+ {
414
+ "epoch": 0.5456760048721072,
415
+ "grad_norm": 0.45864835381507874,
416
+ "learning_rate": 9.567727288213005e-05,
417
+ "loss": 0.5908,
418
+ "step": 56
419
+ },
420
+ {
421
+ "epoch": 0.5554202192448234,
422
+ "grad_norm": 0.4483995735645294,
423
+ "learning_rate": 9.538887392664544e-05,
424
+ "loss": 0.5695,
425
+ "step": 57
426
+ },
427
+ {
428
+ "epoch": 0.5651644336175395,
429
+ "grad_norm": 0.48954787850379944,
430
+ "learning_rate": 9.50916263202557e-05,
431
+ "loss": 0.5314,
432
+ "step": 58
433
+ },
434
+ {
435
+ "epoch": 0.5749086479902558,
436
+ "grad_norm": 0.515805184841156,
437
+ "learning_rate": 9.478558801197065e-05,
438
+ "loss": 0.6366,
439
+ "step": 59
440
+ },
441
+ {
442
+ "epoch": 0.584652862362972,
443
+ "grad_norm": 0.4741595387458801,
444
+ "learning_rate": 9.447081866456489e-05,
445
+ "loss": 0.4933,
446
+ "step": 60
447
+ },
448
+ {
449
+ "epoch": 0.5943970767356882,
450
+ "grad_norm": 0.4829859137535095,
451
+ "learning_rate": 9.414737964294636e-05,
452
+ "loss": 0.6337,
453
+ "step": 61
454
+ },
455
+ {
456
+ "epoch": 0.6041412911084044,
457
+ "grad_norm": 0.45033347606658936,
458
+ "learning_rate": 9.381533400219318e-05,
459
+ "loss": 0.5714,
460
+ "step": 62
461
+ },
462
+ {
463
+ "epoch": 0.6138855054811205,
464
+ "grad_norm": 0.4400843381881714,
465
+ "learning_rate": 9.347474647526095e-05,
466
+ "loss": 0.5756,
467
+ "step": 63
468
+ },
469
+ {
470
+ "epoch": 0.6236297198538368,
471
+ "grad_norm": 0.45498228073120117,
472
+ "learning_rate": 9.312568346036288e-05,
473
+ "loss": 0.5635,
474
+ "step": 64
475
+ },
476
+ {
477
+ "epoch": 0.633373934226553,
478
+ "grad_norm": 0.4397442936897278,
479
+ "learning_rate": 9.276821300802534e-05,
480
+ "loss": 0.5508,
481
+ "step": 65
482
+ },
483
+ {
484
+ "epoch": 0.6431181485992692,
485
+ "grad_norm": 0.465465247631073,
486
+ "learning_rate": 9.24024048078213e-05,
487
+ "loss": 0.5118,
488
+ "step": 66
489
+ },
490
+ {
491
+ "epoch": 0.6528623629719854,
492
+ "grad_norm": 0.4904666244983673,
493
+ "learning_rate": 9.202833017478422e-05,
494
+ "loss": 0.5555,
495
+ "step": 67
496
+ },
497
+ {
498
+ "epoch": 0.6626065773447016,
499
+ "grad_norm": 0.4522901177406311,
500
+ "learning_rate": 9.164606203550497e-05,
501
+ "loss": 0.5712,
502
+ "step": 68
503
+ },
504
+ {
505
+ "epoch": 0.6723507917174177,
506
+ "grad_norm": 0.46516191959381104,
507
+ "learning_rate": 9.125567491391476e-05,
508
+ "loss": 0.54,
509
+ "step": 69
510
+ },
511
+ {
512
+ "epoch": 0.682095006090134,
513
+ "grad_norm": 0.563292384147644,
514
+ "learning_rate": 9.085724491675642e-05,
515
+ "loss": 0.5334,
516
+ "step": 70
517
+ },
518
+ {
519
+ "epoch": 0.6918392204628502,
520
+ "grad_norm": 0.4758321940898895,
521
+ "learning_rate": 9.045084971874738e-05,
522
+ "loss": 0.558,
523
+ "step": 71
524
+ },
525
+ {
526
+ "epoch": 0.7015834348355664,
527
+ "grad_norm": 0.4505232870578766,
528
+ "learning_rate": 9.003656854743667e-05,
529
+ "loss": 0.5592,
530
+ "step": 72
531
+ },
532
+ {
533
+ "epoch": 0.7113276492082826,
534
+ "grad_norm": 0.49200811982154846,
535
+ "learning_rate": 8.961448216775954e-05,
536
+ "loss": 0.562,
537
+ "step": 73
538
+ },
539
+ {
540
+ "epoch": 0.7210718635809987,
541
+ "grad_norm": 0.49411916732788086,
542
+ "learning_rate": 8.9184672866292e-05,
543
+ "loss": 0.5502,
544
+ "step": 74
545
+ },
546
+ {
547
+ "epoch": 0.730816077953715,
548
+ "grad_norm": 0.49506062269210815,
549
+ "learning_rate": 8.874722443520899e-05,
550
+ "loss": 0.5268,
551
+ "step": 75
552
+ },
553
+ {
554
+ "epoch": 0.730816077953715,
555
+ "eval_loss": 0.5484524369239807,
556
+ "eval_runtime": 116.2855,
557
+ "eval_samples_per_second": 7.37,
558
+ "eval_steps_per_second": 0.464,
559
+ "step": 75
560
+ },
561
+ {
562
+ "epoch": 0.7405602923264312,
563
+ "grad_norm": 0.4459410607814789,
564
+ "learning_rate": 8.83022221559489e-05,
565
+ "loss": 0.5651,
566
+ "step": 76
567
+ },
568
+ {
569
+ "epoch": 0.7503045066991474,
570
+ "grad_norm": 0.4649423360824585,
571
+ "learning_rate": 8.784975278258783e-05,
572
+ "loss": 0.507,
573
+ "step": 77
574
+ },
575
+ {
576
+ "epoch": 0.7600487210718636,
577
+ "grad_norm": 0.5267335176467896,
578
+ "learning_rate": 8.73899045249266e-05,
579
+ "loss": 0.5846,
580
+ "step": 78
581
+ },
582
+ {
583
+ "epoch": 0.7697929354445798,
584
+ "grad_norm": 0.48191651701927185,
585
+ "learning_rate": 8.692276703129421e-05,
586
+ "loss": 0.4785,
587
+ "step": 79
588
+ },
589
+ {
590
+ "epoch": 0.7795371498172959,
591
+ "grad_norm": 0.4713154137134552,
592
+ "learning_rate": 8.644843137107059e-05,
593
+ "loss": 0.5115,
594
+ "step": 80
595
+ },
596
+ {
597
+ "epoch": 0.7892813641900122,
598
+ "grad_norm": 0.48336002230644226,
599
+ "learning_rate": 8.596699001693255e-05,
600
+ "loss": 0.4972,
601
+ "step": 81
602
+ },
603
+ {
604
+ "epoch": 0.7990255785627284,
605
+ "grad_norm": 0.5115373730659485,
606
+ "learning_rate": 8.547853682682604e-05,
607
+ "loss": 0.5548,
608
+ "step": 82
609
+ },
610
+ {
611
+ "epoch": 0.8087697929354446,
612
+ "grad_norm": 0.4678809344768524,
613
+ "learning_rate": 8.498316702566828e-05,
614
+ "loss": 0.5017,
615
+ "step": 83
616
+ },
617
+ {
618
+ "epoch": 0.8185140073081608,
619
+ "grad_norm": 0.46973857283592224,
620
+ "learning_rate": 8.44809771867835e-05,
621
+ "loss": 0.5304,
622
+ "step": 84
623
+ },
624
+ {
625
+ "epoch": 0.8282582216808769,
626
+ "grad_norm": 0.5022630095481873,
627
+ "learning_rate": 8.397206521307584e-05,
628
+ "loss": 0.4957,
629
+ "step": 85
630
+ },
631
+ {
632
+ "epoch": 0.8380024360535931,
633
+ "grad_norm": 0.47750282287597656,
634
+ "learning_rate": 8.345653031794292e-05,
635
+ "loss": 0.4853,
636
+ "step": 86
637
+ },
638
+ {
639
+ "epoch": 0.8477466504263094,
640
+ "grad_norm": 0.48647361993789673,
641
+ "learning_rate": 8.293447300593402e-05,
642
+ "loss": 0.499,
643
+ "step": 87
644
+ },
645
+ {
646
+ "epoch": 0.8574908647990256,
647
+ "grad_norm": 0.5240001082420349,
648
+ "learning_rate": 8.240599505315655e-05,
649
+ "loss": 0.5061,
650
+ "step": 88
651
+ },
652
+ {
653
+ "epoch": 0.8672350791717418,
654
+ "grad_norm": 0.4904966354370117,
655
+ "learning_rate": 8.18711994874345e-05,
656
+ "loss": 0.5139,
657
+ "step": 89
658
+ },
659
+ {
660
+ "epoch": 0.876979293544458,
661
+ "grad_norm": 0.5042212605476379,
662
+ "learning_rate": 8.133019056822304e-05,
663
+ "loss": 0.5355,
664
+ "step": 90
665
+ },
666
+ {
667
+ "epoch": 0.8867235079171741,
668
+ "grad_norm": 0.4775819778442383,
669
+ "learning_rate": 8.07830737662829e-05,
670
+ "loss": 0.4889,
671
+ "step": 91
672
+ },
673
+ {
674
+ "epoch": 0.8964677222898904,
675
+ "grad_norm": 0.5047743320465088,
676
+ "learning_rate": 8.022995574311876e-05,
677
+ "loss": 0.4788,
678
+ "step": 92
679
+ },
680
+ {
681
+ "epoch": 0.9062119366626066,
682
+ "grad_norm": 0.4905475676059723,
683
+ "learning_rate": 7.967094433018508e-05,
684
+ "loss": 0.5104,
685
+ "step": 93
686
+ },
687
+ {
688
+ "epoch": 0.9159561510353228,
689
+ "grad_norm": 0.49578583240509033,
690
+ "learning_rate": 7.910614850786448e-05,
691
+ "loss": 0.4903,
692
+ "step": 94
693
+ },
694
+ {
695
+ "epoch": 0.925700365408039,
696
+ "grad_norm": 0.5329849123954773,
697
+ "learning_rate": 7.85356783842216e-05,
698
+ "loss": 0.4926,
699
+ "step": 95
700
+ },
701
+ {
702
+ "epoch": 0.9354445797807551,
703
+ "grad_norm": 0.5267957448959351,
704
+ "learning_rate": 7.795964517353735e-05,
705
+ "loss": 0.5514,
706
+ "step": 96
707
+ },
708
+ {
709
+ "epoch": 0.9451887941534713,
710
+ "grad_norm": 0.5095996856689453,
711
+ "learning_rate": 7.737816117462752e-05,
712
+ "loss": 0.4886,
713
+ "step": 97
714
+ },
715
+ {
716
+ "epoch": 0.9549330085261876,
717
+ "grad_norm": 0.46476784348487854,
718
+ "learning_rate": 7.679133974894983e-05,
719
+ "loss": 0.4776,
720
+ "step": 98
721
+ },
722
+ {
723
+ "epoch": 0.9646772228989038,
724
+ "grad_norm": 0.516237735748291,
725
+ "learning_rate": 7.619929529850397e-05,
726
+ "loss": 0.5343,
727
+ "step": 99
728
+ },
729
+ {
730
+ "epoch": 0.97442143727162,
731
+ "grad_norm": 0.5128530859947205,
732
+ "learning_rate": 7.560214324352858e-05,
733
+ "loss": 0.492,
734
+ "step": 100
735
+ },
736
+ {
737
+ "epoch": 0.97442143727162,
738
+ "eval_loss": 0.5159465074539185,
739
+ "eval_runtime": 116.1788,
740
+ "eval_samples_per_second": 7.377,
741
+ "eval_steps_per_second": 0.465,
742
+ "step": 100
743
+ },
744
+ {
745
+ "epoch": 0.9841656516443362,
746
+ "grad_norm": 0.525909960269928,
747
+ "learning_rate": 7.500000000000001e-05,
748
+ "loss": 0.5276,
749
+ "step": 101
750
+ },
751
+ {
752
+ "epoch": 0.9939098660170523,
753
+ "grad_norm": 0.4755876958370209,
754
+ "learning_rate": 7.439298295693665e-05,
755
+ "loss": 0.4887,
756
+ "step": 102
757
+ },
758
+ {
759
+ "epoch": 1.0097442143727162,
760
+ "grad_norm": 0.898318350315094,
761
+ "learning_rate": 7.378121045351378e-05,
762
+ "loss": 0.896,
763
+ "step": 103
764
+ },
765
+ {
766
+ "epoch": 1.0194884287454324,
767
+ "grad_norm": 0.5060648918151855,
768
+ "learning_rate": 7.316480175599309e-05,
769
+ "loss": 0.4999,
770
+ "step": 104
771
+ },
772
+ {
773
+ "epoch": 1.0292326431181487,
774
+ "grad_norm": 0.46954262256622314,
775
+ "learning_rate": 7.254387703447154e-05,
776
+ "loss": 0.4633,
777
+ "step": 105
778
+ },
779
+ {
780
+ "epoch": 1.0389768574908649,
781
+ "grad_norm": 0.4725915193557739,
782
+ "learning_rate": 7.191855733945387e-05,
783
+ "loss": 0.4491,
784
+ "step": 106
785
+ },
786
+ {
787
+ "epoch": 1.048721071863581,
788
+ "grad_norm": 0.4837670624256134,
789
+ "learning_rate": 7.128896457825364e-05,
790
+ "loss": 0.4508,
791
+ "step": 107
792
+ },
793
+ {
794
+ "epoch": 1.058465286236297,
795
+ "grad_norm": 0.5884471535682678,
796
+ "learning_rate": 7.06552214912271e-05,
797
+ "loss": 0.4297,
798
+ "step": 108
799
+ },
800
+ {
801
+ "epoch": 1.0682095006090133,
802
+ "grad_norm": 0.5307117104530334,
803
+ "learning_rate": 7.001745162784477e-05,
804
+ "loss": 0.4592,
805
+ "step": 109
806
+ },
807
+ {
808
+ "epoch": 1.0779537149817295,
809
+ "grad_norm": 0.5245863795280457,
810
+ "learning_rate": 6.937577932260515e-05,
811
+ "loss": 0.4819,
812
+ "step": 110
813
+ },
814
+ {
815
+ "epoch": 1.0876979293544458,
816
+ "grad_norm": 0.5525639653205872,
817
+ "learning_rate": 6.873032967079561e-05,
818
+ "loss": 0.4678,
819
+ "step": 111
820
+ },
821
+ {
822
+ "epoch": 1.097442143727162,
823
+ "grad_norm": 0.5659409761428833,
824
+ "learning_rate": 6.808122850410461e-05,
825
+ "loss": 0.4635,
826
+ "step": 112
827
+ },
828
+ {
829
+ "epoch": 1.1071863580998782,
830
+ "grad_norm": 0.5493988394737244,
831
+ "learning_rate": 6.742860236609077e-05,
832
+ "loss": 0.4307,
833
+ "step": 113
834
+ },
835
+ {
836
+ "epoch": 1.1169305724725944,
837
+ "grad_norm": 0.5591247081756592,
838
+ "learning_rate": 6.677257848751277e-05,
839
+ "loss": 0.416,
840
+ "step": 114
841
+ },
842
+ {
843
+ "epoch": 1.1266747868453106,
844
+ "grad_norm": 0.4955357611179352,
845
+ "learning_rate": 6.611328476152557e-05,
846
+ "loss": 0.4662,
847
+ "step": 115
848
+ },
849
+ {
850
+ "epoch": 1.1364190012180269,
851
+ "grad_norm": 0.5241750478744507,
852
+ "learning_rate": 6.545084971874738e-05,
853
+ "loss": 0.4301,
854
+ "step": 116
855
+ },
856
+ {
857
+ "epoch": 1.146163215590743,
858
+ "grad_norm": 0.4876319468021393,
859
+ "learning_rate": 6.478540250220234e-05,
860
+ "loss": 0.4564,
861
+ "step": 117
862
+ },
863
+ {
864
+ "epoch": 1.155907429963459,
865
+ "grad_norm": 0.5395438075065613,
866
+ "learning_rate": 6.411707284214384e-05,
867
+ "loss": 0.502,
868
+ "step": 118
869
+ },
870
+ {
871
+ "epoch": 1.1656516443361755,
872
+ "grad_norm": 0.5435388684272766,
873
+ "learning_rate": 6.344599103076329e-05,
874
+ "loss": 0.4536,
875
+ "step": 119
876
+ },
877
+ {
878
+ "epoch": 1.1753958587088915,
879
+ "grad_norm": 0.526055097579956,
880
+ "learning_rate": 6.277228789678953e-05,
881
+ "loss": 0.4576,
882
+ "step": 120
883
+ },
884
+ {
885
+ "epoch": 1.1851400730816077,
886
+ "grad_norm": 0.5060620903968811,
887
+ "learning_rate": 6.209609477998338e-05,
888
+ "loss": 0.4814,
889
+ "step": 121
890
+ },
891
+ {
892
+ "epoch": 1.194884287454324,
893
+ "grad_norm": 0.4862349033355713,
894
+ "learning_rate": 6.141754350553279e-05,
895
+ "loss": 0.4277,
896
+ "step": 122
897
+ },
898
+ {
899
+ "epoch": 1.2046285018270402,
900
+ "grad_norm": 0.5572500824928284,
901
+ "learning_rate": 6.073676635835317e-05,
902
+ "loss": 0.5118,
903
+ "step": 123
904
+ },
905
+ {
906
+ "epoch": 1.2143727161997564,
907
+ "grad_norm": 0.5726194381713867,
908
+ "learning_rate": 6.005389605729824e-05,
909
+ "loss": 0.4725,
910
+ "step": 124
911
+ },
912
+ {
913
+ "epoch": 1.2241169305724726,
914
+ "grad_norm": 0.5314717888832092,
915
+ "learning_rate": 5.9369065729286245e-05,
916
+ "loss": 0.4719,
917
+ "step": 125
918
+ },
919
+ {
920
+ "epoch": 1.2241169305724726,
921
+ "eval_loss": 0.4989548325538635,
922
+ "eval_runtime": 116.3548,
923
+ "eval_samples_per_second": 7.365,
924
+ "eval_steps_per_second": 0.464,
925
+ "step": 125
926
+ },
927
+ {
928
+ "epoch": 1.2338611449451888,
929
+ "grad_norm": 0.5402824878692627,
930
+ "learning_rate": 5.868240888334653e-05,
931
+ "loss": 0.4735,
932
+ "step": 126
933
+ },
934
+ {
935
+ "epoch": 1.243605359317905,
936
+ "grad_norm": 0.5422326922416687,
937
+ "learning_rate": 5.799405938459175e-05,
938
+ "loss": 0.4729,
939
+ "step": 127
940
+ },
941
+ {
942
+ "epoch": 1.2533495736906213,
943
+ "grad_norm": 0.5375432968139648,
944
+ "learning_rate": 5.730415142812059e-05,
945
+ "loss": 0.4589,
946
+ "step": 128
947
+ },
948
+ {
949
+ "epoch": 1.2630937880633373,
950
+ "grad_norm": 0.5534482002258301,
951
+ "learning_rate": 5.661281951285613e-05,
952
+ "loss": 0.4464,
953
+ "step": 129
954
+ },
955
+ {
956
+ "epoch": 1.2728380024360537,
957
+ "grad_norm": 0.5432469844818115,
958
+ "learning_rate": 5.5920198415325064e-05,
959
+ "loss": 0.4537,
960
+ "step": 130
961
+ },
962
+ {
963
+ "epoch": 1.2825822168087697,
964
+ "grad_norm": 0.5186154246330261,
965
+ "learning_rate": 5.522642316338268e-05,
966
+ "loss": 0.4158,
967
+ "step": 131
968
+ },
969
+ {
970
+ "epoch": 1.292326431181486,
971
+ "grad_norm": 0.5382589101791382,
972
+ "learning_rate": 5.453162900988902e-05,
973
+ "loss": 0.4176,
974
+ "step": 132
975
+ },
976
+ {
977
+ "epoch": 1.3020706455542022,
978
+ "grad_norm": 0.5565094351768494,
979
+ "learning_rate": 5.383595140634093e-05,
980
+ "loss": 0.4287,
981
+ "step": 133
982
+ },
983
+ {
984
+ "epoch": 1.3118148599269184,
985
+ "grad_norm": 0.5403843522071838,
986
+ "learning_rate": 5.313952597646568e-05,
987
+ "loss": 0.4118,
988
+ "step": 134
989
+ },
990
+ {
991
+ "epoch": 1.3215590742996346,
992
+ "grad_norm": 0.5338913798332214,
993
+ "learning_rate": 5.244248848978067e-05,
994
+ "loss": 0.3948,
995
+ "step": 135
996
+ },
997
+ {
998
+ "epoch": 1.3313032886723508,
999
+ "grad_norm": 0.5245387554168701,
1000
+ "learning_rate": 5.174497483512506e-05,
1001
+ "loss": 0.4579,
1002
+ "step": 136
1003
+ },
1004
+ {
1005
+ "epoch": 1.341047503045067,
1006
+ "grad_norm": 0.51710045337677,
1007
+ "learning_rate": 5.104712099416785e-05,
1008
+ "loss": 0.4025,
1009
+ "step": 137
1010
+ },
1011
+ {
1012
+ "epoch": 1.3507917174177833,
1013
+ "grad_norm": 0.56063312292099,
1014
+ "learning_rate": 5.034906301489808e-05,
1015
+ "loss": 0.4956,
1016
+ "step": 138
1017
+ },
1018
+ {
1019
+ "epoch": 1.3605359317904995,
1020
+ "grad_norm": 0.6041215658187866,
1021
+ "learning_rate": 4.965093698510193e-05,
1022
+ "loss": 0.4336,
1023
+ "step": 139
1024
+ },
1025
+ {
1026
+ "epoch": 1.3702801461632155,
1027
+ "grad_norm": 0.5541957020759583,
1028
+ "learning_rate": 4.895287900583216e-05,
1029
+ "loss": 0.4456,
1030
+ "step": 140
1031
+ },
1032
+ {
1033
+ "epoch": 1.380024360535932,
1034
+ "grad_norm": 0.5618797540664673,
1035
+ "learning_rate": 4.825502516487497e-05,
1036
+ "loss": 0.3996,
1037
+ "step": 141
1038
+ },
1039
+ {
1040
+ "epoch": 1.389768574908648,
1041
+ "grad_norm": 0.533968448638916,
1042
+ "learning_rate": 4.755751151021934e-05,
1043
+ "loss": 0.4155,
1044
+ "step": 142
1045
+ },
1046
+ {
1047
+ "epoch": 1.3995127892813641,
1048
+ "grad_norm": 0.5411080121994019,
1049
+ "learning_rate": 4.6860474023534335e-05,
1050
+ "loss": 0.4325,
1051
+ "step": 143
1052
+ },
1053
+ {
1054
+ "epoch": 1.4092570036540804,
1055
+ "grad_norm": 0.5761292576789856,
1056
+ "learning_rate": 4.616404859365907e-05,
1057
+ "loss": 0.4166,
1058
+ "step": 144
1059
+ },
1060
+ {
1061
+ "epoch": 1.4190012180267966,
1062
+ "grad_norm": 0.5804659128189087,
1063
+ "learning_rate": 4.5468370990111006e-05,
1064
+ "loss": 0.4436,
1065
+ "step": 145
1066
+ },
1067
+ {
1068
+ "epoch": 1.4287454323995128,
1069
+ "grad_norm": 0.5537294745445251,
1070
+ "learning_rate": 4.477357683661734e-05,
1071
+ "loss": 0.3795,
1072
+ "step": 146
1073
+ },
1074
+ {
1075
+ "epoch": 1.438489646772229,
1076
+ "grad_norm": 0.5477214455604553,
1077
+ "learning_rate": 4.407980158467495e-05,
1078
+ "loss": 0.3896,
1079
+ "step": 147
1080
+ },
1081
+ {
1082
+ "epoch": 1.4482338611449452,
1083
+ "grad_norm": 0.5110722780227661,
1084
+ "learning_rate": 4.3387180487143876e-05,
1085
+ "loss": 0.4036,
1086
+ "step": 148
1087
+ },
1088
+ {
1089
+ "epoch": 1.4579780755176615,
1090
+ "grad_norm": 0.5480209589004517,
1091
+ "learning_rate": 4.269584857187943e-05,
1092
+ "loss": 0.4451,
1093
+ "step": 149
1094
+ },
1095
+ {
1096
+ "epoch": 1.4677222898903777,
1097
+ "grad_norm": 0.5255818367004395,
1098
+ "learning_rate": 4.2005940615408264e-05,
1099
+ "loss": 0.4031,
1100
+ "step": 150
1101
+ },
1102
+ {
1103
+ "epoch": 1.4677222898903777,
1104
+ "eval_loss": 0.4808570146560669,
1105
+ "eval_runtime": 116.3103,
1106
+ "eval_samples_per_second": 7.368,
1107
+ "eval_steps_per_second": 0.464,
1108
+ "step": 150
1109
+ },
1110
+ {
1111
+ "epoch": 1.4774665042630937,
1112
+ "grad_norm": 0.5485397577285767,
1113
+ "learning_rate": 4.131759111665349e-05,
1114
+ "loss": 0.462,
1115
+ "step": 151
1116
+ },
1117
+ {
1118
+ "epoch": 1.4872107186358101,
1119
+ "grad_norm": 0.5463838577270508,
1120
+ "learning_rate": 4.063093427071376e-05,
1121
+ "loss": 0.4265,
1122
+ "step": 152
1123
+ },
1124
+ {
1125
+ "epoch": 1.4969549330085261,
1126
+ "grad_norm": 0.5511963367462158,
1127
+ "learning_rate": 3.9946103942701777e-05,
1128
+ "loss": 0.3964,
1129
+ "step": 153
1130
+ },
1131
+ {
1132
+ "epoch": 1.5066991473812423,
1133
+ "grad_norm": 0.5571511387825012,
1134
+ "learning_rate": 3.926323364164684e-05,
1135
+ "loss": 0.4642,
1136
+ "step": 154
1137
+ },
1138
+ {
1139
+ "epoch": 1.5164433617539586,
1140
+ "grad_norm": 0.5894230604171753,
1141
+ "learning_rate": 3.858245649446721e-05,
1142
+ "loss": 0.4511,
1143
+ "step": 155
1144
+ },
1145
+ {
1146
+ "epoch": 1.5261875761266748,
1147
+ "grad_norm": 0.5373761653900146,
1148
+ "learning_rate": 3.790390522001662e-05,
1149
+ "loss": 0.388,
1150
+ "step": 156
1151
+ },
1152
+ {
1153
+ "epoch": 1.535931790499391,
1154
+ "grad_norm": 0.5737512707710266,
1155
+ "learning_rate": 3.7227712103210486e-05,
1156
+ "loss": 0.375,
1157
+ "step": 157
1158
+ },
1159
+ {
1160
+ "epoch": 1.5456760048721072,
1161
+ "grad_norm": 0.5502772331237793,
1162
+ "learning_rate": 3.655400896923672e-05,
1163
+ "loss": 0.4433,
1164
+ "step": 158
1165
+ },
1166
+ {
1167
+ "epoch": 1.5554202192448234,
1168
+ "grad_norm": 0.5434299111366272,
1169
+ "learning_rate": 3.588292715785617e-05,
1170
+ "loss": 0.4002,
1171
+ "step": 159
1172
+ },
1173
+ {
1174
+ "epoch": 1.5651644336175394,
1175
+ "grad_norm": 0.5924922227859497,
1176
+ "learning_rate": 3.5214597497797684e-05,
1177
+ "loss": 0.4141,
1178
+ "step": 160
1179
+ },
1180
+ {
1181
+ "epoch": 1.5749086479902559,
1182
+ "grad_norm": 0.6273884177207947,
1183
+ "learning_rate": 3.4549150281252636e-05,
1184
+ "loss": 0.4277,
1185
+ "step": 161
1186
+ },
1187
+ {
1188
+ "epoch": 1.5846528623629719,
1189
+ "grad_norm": 0.578081488609314,
1190
+ "learning_rate": 3.388671523847445e-05,
1191
+ "loss": 0.3956,
1192
+ "step": 162
1193
+ },
1194
+ {
1195
+ "epoch": 1.5943970767356883,
1196
+ "grad_norm": 0.5401508212089539,
1197
+ "learning_rate": 3.322742151248725e-05,
1198
+ "loss": 0.4305,
1199
+ "step": 163
1200
+ },
1201
+ {
1202
+ "epoch": 1.6041412911084043,
1203
+ "grad_norm": 0.6144226789474487,
1204
+ "learning_rate": 3.257139763390925e-05,
1205
+ "loss": 0.4126,
1206
+ "step": 164
1207
+ },
1208
+ {
1209
+ "epoch": 1.6138855054811205,
1210
+ "grad_norm": 0.5739960670471191,
1211
+ "learning_rate": 3.1918771495895396e-05,
1212
+ "loss": 0.404,
1213
+ "step": 165
1214
+ },
1215
+ {
1216
+ "epoch": 1.6236297198538368,
1217
+ "grad_norm": 0.5783050060272217,
1218
+ "learning_rate": 3.12696703292044e-05,
1219
+ "loss": 0.4209,
1220
+ "step": 166
1221
+ },
1222
+ {
1223
+ "epoch": 1.633373934226553,
1224
+ "grad_norm": 0.5928875207901001,
1225
+ "learning_rate": 3.062422067739485e-05,
1226
+ "loss": 0.4317,
1227
+ "step": 167
1228
+ },
1229
+ {
1230
+ "epoch": 1.6431181485992692,
1231
+ "grad_norm": 0.5249680876731873,
1232
+ "learning_rate": 2.9982548372155263e-05,
1233
+ "loss": 0.3964,
1234
+ "step": 168
1235
+ },
1236
+ {
1237
+ "epoch": 1.6528623629719854,
1238
+ "grad_norm": 0.5612180233001709,
1239
+ "learning_rate": 2.934477850877292e-05,
1240
+ "loss": 0.3919,
1241
+ "step": 169
1242
+ },
1243
+ {
1244
+ "epoch": 1.6626065773447016,
1245
+ "grad_norm": 0.5700953006744385,
1246
+ "learning_rate": 2.8711035421746367e-05,
1247
+ "loss": 0.4403,
1248
+ "step": 170
1249
+ },
1250
+ {
1251
+ "epoch": 1.6723507917174176,
1252
+ "grad_norm": 0.5822266936302185,
1253
+ "learning_rate": 2.8081442660546125e-05,
1254
+ "loss": 0.4265,
1255
+ "step": 171
1256
+ },
1257
+ {
1258
+ "epoch": 1.682095006090134,
1259
+ "grad_norm": 0.5738089680671692,
1260
+ "learning_rate": 2.7456122965528475e-05,
1261
+ "loss": 0.4612,
1262
+ "step": 172
1263
+ },
1264
+ {
1265
+ "epoch": 1.69183922046285,
1266
+ "grad_norm": 0.5751779079437256,
1267
+ "learning_rate": 2.6835198244006927e-05,
1268
+ "loss": 0.3983,
1269
+ "step": 173
1270
+ },
1271
+ {
1272
+ "epoch": 1.7015834348355665,
1273
+ "grad_norm": 0.5598529577255249,
1274
+ "learning_rate": 2.6218789546486234e-05,
1275
+ "loss": 0.4338,
1276
+ "step": 174
1277
+ },
1278
+ {
1279
+ "epoch": 1.7113276492082825,
1280
+ "grad_norm": 0.5587442517280579,
1281
+ "learning_rate": 2.560701704306336e-05,
1282
+ "loss": 0.3825,
1283
+ "step": 175
1284
+ },
1285
+ {
1286
+ "epoch": 1.7113276492082825,
1287
+ "eval_loss": 0.47023555636405945,
1288
+ "eval_runtime": 116.2392,
1289
+ "eval_samples_per_second": 7.373,
1290
+ "eval_steps_per_second": 0.465,
1291
+ "step": 175
1292
+ },
1293
+ {
1294
+ "epoch": 1.7210718635809987,
1295
+ "grad_norm": 0.5803806185722351,
1296
+ "learning_rate": 2.500000000000001e-05,
1297
+ "loss": 0.4265,
1298
+ "step": 176
1299
+ },
1300
+ {
1301
+ "epoch": 1.730816077953715,
1302
+ "grad_norm": 0.5986764430999756,
1303
+ "learning_rate": 2.4397856756471432e-05,
1304
+ "loss": 0.4299,
1305
+ "step": 177
1306
+ },
1307
+ {
1308
+ "epoch": 1.7405602923264312,
1309
+ "grad_norm": 0.6137004494667053,
1310
+ "learning_rate": 2.3800704701496053e-05,
1311
+ "loss": 0.4355,
1312
+ "step": 178
1313
+ },
1314
+ {
1315
+ "epoch": 1.7503045066991474,
1316
+ "grad_norm": 0.5777759552001953,
1317
+ "learning_rate": 2.3208660251050158e-05,
1318
+ "loss": 0.4235,
1319
+ "step": 179
1320
+ },
1321
+ {
1322
+ "epoch": 1.7600487210718636,
1323
+ "grad_norm": 0.599251389503479,
1324
+ "learning_rate": 2.2621838825372493e-05,
1325
+ "loss": 0.4078,
1326
+ "step": 180
1327
+ },
1328
+ {
1329
+ "epoch": 1.7697929354445798,
1330
+ "grad_norm": 0.5639720559120178,
1331
+ "learning_rate": 2.2040354826462668e-05,
1332
+ "loss": 0.4046,
1333
+ "step": 181
1334
+ },
1335
+ {
1336
+ "epoch": 1.7795371498172958,
1337
+ "grad_norm": 0.5639522075653076,
1338
+ "learning_rate": 2.1464321615778422e-05,
1339
+ "loss": 0.433,
1340
+ "step": 182
1341
+ },
1342
+ {
1343
+ "epoch": 1.7892813641900123,
1344
+ "grad_norm": 0.5366583466529846,
1345
+ "learning_rate": 2.0893851492135537e-05,
1346
+ "loss": 0.4318,
1347
+ "step": 183
1348
+ },
1349
+ {
1350
+ "epoch": 1.7990255785627283,
1351
+ "grad_norm": 0.5910248160362244,
1352
+ "learning_rate": 2.0329055669814934e-05,
1353
+ "loss": 0.4184,
1354
+ "step": 184
1355
+ },
1356
+ {
1357
+ "epoch": 1.8087697929354447,
1358
+ "grad_norm": 0.5742807388305664,
1359
+ "learning_rate": 1.977004425688126e-05,
1360
+ "loss": 0.4346,
1361
+ "step": 185
1362
+ },
1363
+ {
1364
+ "epoch": 1.8185140073081607,
1365
+ "grad_norm": 0.5683897137641907,
1366
+ "learning_rate": 1.9216926233717085e-05,
1367
+ "loss": 0.4007,
1368
+ "step": 186
1369
+ },
1370
+ {
1371
+ "epoch": 1.828258221680877,
1372
+ "grad_norm": 0.5249460339546204,
1373
+ "learning_rate": 1.866980943177699e-05,
1374
+ "loss": 0.4165,
1375
+ "step": 187
1376
+ },
1377
+ {
1378
+ "epoch": 1.8380024360535931,
1379
+ "grad_norm": 0.5429449677467346,
1380
+ "learning_rate": 1.8128800512565513e-05,
1381
+ "loss": 0.378,
1382
+ "step": 188
1383
+ },
1384
+ {
1385
+ "epoch": 1.8477466504263094,
1386
+ "grad_norm": 0.5251733660697937,
1387
+ "learning_rate": 1.7594004946843456e-05,
1388
+ "loss": 0.3824,
1389
+ "step": 189
1390
+ },
1391
+ {
1392
+ "epoch": 1.8574908647990256,
1393
+ "grad_norm": 0.5763549208641052,
1394
+ "learning_rate": 1.7065526994065973e-05,
1395
+ "loss": 0.3788,
1396
+ "step": 190
1397
+ },
1398
+ {
1399
+ "epoch": 1.8672350791717418,
1400
+ "grad_norm": 0.5946308970451355,
1401
+ "learning_rate": 1.6543469682057106e-05,
1402
+ "loss": 0.448,
1403
+ "step": 191
1404
+ },
1405
+ {
1406
+ "epoch": 1.876979293544458,
1407
+ "grad_norm": 0.5707454085350037,
1408
+ "learning_rate": 1.602793478692419e-05,
1409
+ "loss": 0.3816,
1410
+ "step": 192
1411
+ },
1412
+ {
1413
+ "epoch": 1.886723507917174,
1414
+ "grad_norm": 0.5635607838630676,
1415
+ "learning_rate": 1.551902281321651e-05,
1416
+ "loss": 0.3979,
1417
+ "step": 193
1418
+ },
1419
+ {
1420
+ "epoch": 1.8964677222898905,
1421
+ "grad_norm": 1.6317293643951416,
1422
+ "learning_rate": 1.5016832974331724e-05,
1423
+ "loss": 0.3654,
1424
+ "step": 194
1425
+ },
1426
+ {
1427
+ "epoch": 1.9062119366626065,
1428
+ "grad_norm": 0.5927464365959167,
1429
+ "learning_rate": 1.4521463173173965e-05,
1430
+ "loss": 0.4432,
1431
+ "step": 195
1432
+ },
1433
+ {
1434
+ "epoch": 1.915956151035323,
1435
+ "grad_norm": 0.5671654343605042,
1436
+ "learning_rate": 1.4033009983067452e-05,
1437
+ "loss": 0.364,
1438
+ "step": 196
1439
+ },
1440
+ {
1441
+ "epoch": 1.925700365408039,
1442
+ "grad_norm": 0.5950899124145508,
1443
+ "learning_rate": 1.3551568628929434e-05,
1444
+ "loss": 0.4442,
1445
+ "step": 197
1446
+ },
1447
+ {
1448
+ "epoch": 1.9354445797807551,
1449
+ "grad_norm": 0.5787199139595032,
1450
+ "learning_rate": 1.3077232968705805e-05,
1451
+ "loss": 0.4042,
1452
+ "step": 198
1453
+ },
1454
+ {
1455
+ "epoch": 1.9451887941534713,
1456
+ "grad_norm": 0.5602390766143799,
1457
+ "learning_rate": 1.2610095475073414e-05,
1458
+ "loss": 0.4252,
1459
+ "step": 199
1460
+ },
1461
+ {
1462
+ "epoch": 1.9549330085261876,
1463
+ "grad_norm": 0.5912858843803406,
1464
+ "learning_rate": 1.2150247217412186e-05,
1465
+ "loss": 0.3944,
1466
+ "step": 200
1467
+ },
1468
+ {
1469
+ "epoch": 1.9549330085261876,
1470
+ "eval_loss": 0.4625219404697418,
1471
+ "eval_runtime": 116.3528,
1472
+ "eval_samples_per_second": 7.366,
1473
+ "eval_steps_per_second": 0.464,
1474
+ "step": 200
1475
+ },
1476
+ {
1477
+ "epoch": 1.9646772228989038,
1478
+ "grad_norm": 0.5247693657875061,
1479
+ "learning_rate": 1.1697777844051105e-05,
1480
+ "loss": 0.397,
1481
+ "step": 201
1482
+ },
1483
+ {
1484
+ "epoch": 1.97442143727162,
1485
+ "grad_norm": 0.5671912431716919,
1486
+ "learning_rate": 1.1252775564791024e-05,
1487
+ "loss": 0.4191,
1488
+ "step": 202
1489
+ },
1490
+ {
1491
+ "epoch": 1.9841656516443362,
1492
+ "grad_norm": 0.5699170827865601,
1493
+ "learning_rate": 1.0815327133708015e-05,
1494
+ "loss": 0.4126,
1495
+ "step": 203
1496
+ },
1497
+ {
1498
+ "epoch": 1.9939098660170522,
1499
+ "grad_norm": 0.5686872005462646,
1500
+ "learning_rate": 1.0385517832240471e-05,
1501
+ "loss": 0.3936,
1502
+ "step": 204
1503
+ },
1504
+ {
1505
+ "epoch": 2.009744214372716,
1506
+ "grad_norm": 0.9932243824005127,
1507
+ "learning_rate": 9.963431452563332e-06,
1508
+ "loss": 0.8769,
1509
+ "step": 205
1510
+ },
1511
+ {
1512
+ "epoch": 2.0194884287454324,
1513
+ "grad_norm": 0.5336227416992188,
1514
+ "learning_rate": 9.549150281252633e-06,
1515
+ "loss": 0.3756,
1516
+ "step": 206
1517
+ },
1518
+ {
1519
+ "epoch": 2.0292326431181484,
1520
+ "grad_norm": 0.5651218891143799,
1521
+ "learning_rate": 9.142755083243576e-06,
1522
+ "loss": 0.4197,
1523
+ "step": 207
1524
+ },
1525
+ {
1526
+ "epoch": 2.038976857490865,
1527
+ "grad_norm": 0.5759946703910828,
1528
+ "learning_rate": 8.744325086085248e-06,
1529
+ "loss": 0.4043,
1530
+ "step": 208
1531
+ },
1532
+ {
1533
+ "epoch": 2.048721071863581,
1534
+ "grad_norm": 0.529643177986145,
1535
+ "learning_rate": 8.353937964495029e-06,
1536
+ "loss": 0.3806,
1537
+ "step": 209
1538
+ },
1539
+ {
1540
+ "epoch": 2.0584652862362973,
1541
+ "grad_norm": 0.563081681728363,
1542
+ "learning_rate": 7.971669825215788e-06,
1543
+ "loss": 0.3838,
1544
+ "step": 210
1545
+ },
1546
+ {
1547
+ "epoch": 2.0682095006090133,
1548
+ "grad_norm": 0.5327872037887573,
1549
+ "learning_rate": 7.597595192178702e-06,
1550
+ "loss": 0.3778,
1551
+ "step": 211
1552
+ },
1553
+ {
1554
+ "epoch": 2.0779537149817298,
1555
+ "grad_norm": 0.5525795817375183,
1556
+ "learning_rate": 7.2317869919746705e-06,
1557
+ "loss": 0.4119,
1558
+ "step": 212
1559
+ },
1560
+ {
1561
+ "epoch": 2.0876979293544458,
1562
+ "grad_norm": 0.5779188871383667,
1563
+ "learning_rate": 6.874316539637127e-06,
1564
+ "loss": 0.3804,
1565
+ "step": 213
1566
+ },
1567
+ {
1568
+ "epoch": 2.097442143727162,
1569
+ "grad_norm": 0.5842592716217041,
1570
+ "learning_rate": 6.52525352473905e-06,
1571
+ "loss": 0.3932,
1572
+ "step": 214
1573
+ },
1574
+ {
1575
+ "epoch": 2.107186358099878,
1576
+ "grad_norm": 0.5471872091293335,
1577
+ "learning_rate": 6.184665997806832e-06,
1578
+ "loss": 0.3658,
1579
+ "step": 215
1580
+ },
1581
+ {
1582
+ "epoch": 2.116930572472594,
1583
+ "grad_norm": 0.5467835664749146,
1584
+ "learning_rate": 5.852620357053651e-06,
1585
+ "loss": 0.3683,
1586
+ "step": 216
1587
+ },
1588
+ {
1589
+ "epoch": 2.1266747868453106,
1590
+ "grad_norm": 0.5607587099075317,
1591
+ "learning_rate": 5.529181335435124e-06,
1592
+ "loss": 0.3635,
1593
+ "step": 217
1594
+ },
1595
+ {
1596
+ "epoch": 2.1364190012180266,
1597
+ "grad_norm": 0.5682641267776489,
1598
+ "learning_rate": 5.214411988029355e-06,
1599
+ "loss": 0.3929,
1600
+ "step": 218
1601
+ },
1602
+ {
1603
+ "epoch": 2.146163215590743,
1604
+ "grad_norm": 0.5493900775909424,
1605
+ "learning_rate": 4.908373679744316e-06,
1606
+ "loss": 0.3579,
1607
+ "step": 219
1608
+ },
1609
+ {
1610
+ "epoch": 2.155907429963459,
1611
+ "grad_norm": 0.5399561524391174,
1612
+ "learning_rate": 4.611126073354571e-06,
1613
+ "loss": 0.397,
1614
+ "step": 220
1615
+ },
1616
+ {
1617
+ "epoch": 2.1656516443361755,
1618
+ "grad_norm": 0.566953182220459,
1619
+ "learning_rate": 4.322727117869951e-06,
1620
+ "loss": 0.3996,
1621
+ "step": 221
1622
+ },
1623
+ {
1624
+ "epoch": 2.1753958587088915,
1625
+ "grad_norm": 0.5458848476409912,
1626
+ "learning_rate": 4.043233037238281e-06,
1627
+ "loss": 0.3797,
1628
+ "step": 222
1629
+ },
1630
+ {
1631
+ "epoch": 2.185140073081608,
1632
+ "grad_norm": 0.5842291116714478,
1633
+ "learning_rate": 3.772698319384349e-06,
1634
+ "loss": 0.4177,
1635
+ "step": 223
1636
+ },
1637
+ {
1638
+ "epoch": 2.194884287454324,
1639
+ "grad_norm": 0.5811557769775391,
1640
+ "learning_rate": 3.511175705587433e-06,
1641
+ "loss": 0.4057,
1642
+ "step": 224
1643
+ },
1644
+ {
1645
+ "epoch": 2.2046285018270404,
1646
+ "grad_norm": 0.5538292527198792,
1647
+ "learning_rate": 3.258716180199278e-06,
1648
+ "loss": 0.3411,
1649
+ "step": 225
1650
+ },
1651
+ {
1652
+ "epoch": 2.2046285018270404,
1653
+ "eval_loss": 0.461640328168869,
1654
+ "eval_runtime": 116.2914,
1655
+ "eval_samples_per_second": 7.369,
1656
+ "eval_steps_per_second": 0.464,
1657
+ "step": 225
1658
+ },
1659
+ {
1660
+ "epoch": 2.2143727161997564,
1661
+ "grad_norm": 0.5587625503540039,
1662
+ "learning_rate": 3.0153689607045845e-06,
1663
+ "loss": 0.3767,
1664
+ "step": 226
1665
+ },
1666
+ {
1667
+ "epoch": 2.2241169305724724,
1668
+ "grad_norm": 0.6179119348526001,
1669
+ "learning_rate": 2.7811814881259503e-06,
1670
+ "loss": 0.3909,
1671
+ "step": 227
1672
+ },
1673
+ {
1674
+ "epoch": 2.233861144945189,
1675
+ "grad_norm": 0.5846551060676575,
1676
+ "learning_rate": 2.5561994177751737e-06,
1677
+ "loss": 0.377,
1678
+ "step": 228
1679
+ },
1680
+ {
1681
+ "epoch": 2.243605359317905,
1682
+ "grad_norm": 0.5826045274734497,
1683
+ "learning_rate": 2.340466610352654e-06,
1684
+ "loss": 0.3621,
1685
+ "step": 229
1686
+ },
1687
+ {
1688
+ "epoch": 2.2533495736906213,
1689
+ "grad_norm": 0.5827783942222595,
1690
+ "learning_rate": 2.134025123396638e-06,
1691
+ "loss": 0.4011,
1692
+ "step": 230
1693
+ },
1694
+ {
1695
+ "epoch": 2.2630937880633373,
1696
+ "grad_norm": 0.5747119188308716,
1697
+ "learning_rate": 1.9369152030840556e-06,
1698
+ "loss": 0.3517,
1699
+ "step": 231
1700
+ },
1701
+ {
1702
+ "epoch": 2.2728380024360537,
1703
+ "grad_norm": 0.5735805630683899,
1704
+ "learning_rate": 1.7491752763844293e-06,
1705
+ "loss": 0.3749,
1706
+ "step": 232
1707
+ },
1708
+ {
1709
+ "epoch": 2.2825822168087697,
1710
+ "grad_norm": 0.5682513117790222,
1711
+ "learning_rate": 1.5708419435684462e-06,
1712
+ "loss": 0.3629,
1713
+ "step": 233
1714
+ },
1715
+ {
1716
+ "epoch": 2.292326431181486,
1717
+ "grad_norm": 0.587968111038208,
1718
+ "learning_rate": 1.4019499710726913e-06,
1719
+ "loss": 0.3971,
1720
+ "step": 234
1721
+ },
1722
+ {
1723
+ "epoch": 2.302070645554202,
1724
+ "grad_norm": 0.5531240105628967,
1725
+ "learning_rate": 1.2425322847218368e-06,
1726
+ "loss": 0.3643,
1727
+ "step": 235
1728
+ },
1729
+ {
1730
+ "epoch": 2.311814859926918,
1731
+ "grad_norm": 0.5963295698165894,
1732
+ "learning_rate": 1.0926199633097157e-06,
1733
+ "loss": 0.3678,
1734
+ "step": 236
1735
+ },
1736
+ {
1737
+ "epoch": 2.3215590742996346,
1738
+ "grad_norm": 0.593429684638977,
1739
+ "learning_rate": 9.522422325404235e-07,
1740
+ "loss": 0.4407,
1741
+ "step": 237
1742
+ },
1743
+ {
1744
+ "epoch": 2.331303288672351,
1745
+ "grad_norm": 0.5391296148300171,
1746
+ "learning_rate": 8.214264593307098e-07,
1747
+ "loss": 0.3228,
1748
+ "step": 238
1749
+ },
1750
+ {
1751
+ "epoch": 2.341047503045067,
1752
+ "grad_norm": 0.5678541660308838,
1753
+ "learning_rate": 7.001981464747565e-07,
1754
+ "loss": 0.3825,
1755
+ "step": 239
1756
+ },
1757
+ {
1758
+ "epoch": 2.350791717417783,
1759
+ "grad_norm": 0.5584573149681091,
1760
+ "learning_rate": 5.885809276723608e-07,
1761
+ "loss": 0.3306,
1762
+ "step": 240
1763
+ },
1764
+ {
1765
+ "epoch": 2.3605359317904995,
1766
+ "grad_norm": 0.5642800331115723,
1767
+ "learning_rate": 4.865965629214819e-07,
1768
+ "loss": 0.368,
1769
+ "step": 241
1770
+ },
1771
+ {
1772
+ "epoch": 2.3702801461632155,
1773
+ "grad_norm": 0.5758326053619385,
1774
+ "learning_rate": 3.9426493427611177e-07,
1775
+ "loss": 0.3755,
1776
+ "step": 242
1777
+ },
1778
+ {
1779
+ "epoch": 2.380024360535932,
1780
+ "grad_norm": 0.5884226560592651,
1781
+ "learning_rate": 3.1160404197018154e-07,
1782
+ "loss": 0.389,
1783
+ "step": 243
1784
+ },
1785
+ {
1786
+ "epoch": 2.389768574908648,
1787
+ "grad_norm": 0.5593277812004089,
1788
+ "learning_rate": 2.386300009084408e-07,
1789
+ "loss": 0.3565,
1790
+ "step": 244
1791
+ },
1792
+ {
1793
+ "epoch": 2.3995127892813644,
1794
+ "grad_norm": 0.5782268643379211,
1795
+ "learning_rate": 1.753570375247815e-07,
1796
+ "loss": 0.3858,
1797
+ "step": 245
1798
+ },
1799
+ {
1800
+ "epoch": 2.4092570036540804,
1801
+ "grad_norm": 0.5798349380493164,
1802
+ "learning_rate": 1.2179748700879012e-07,
1803
+ "loss": 0.3403,
1804
+ "step": 246
1805
+ },
1806
+ {
1807
+ "epoch": 2.4190012180267964,
1808
+ "grad_norm": 0.596785843372345,
1809
+ "learning_rate": 7.796179090094891e-08,
1810
+ "loss": 0.3451,
1811
+ "step": 247
1812
+ },
1813
+ {
1814
+ "epoch": 2.428745432399513,
1815
+ "grad_norm": 0.5822293162345886,
1816
+ "learning_rate": 4.385849505708084e-08,
1817
+ "loss": 0.3843,
1818
+ "step": 248
1819
+ },
1820
+ {
1821
+ "epoch": 2.4384896467722292,
1822
+ "grad_norm": 0.5595471262931824,
1823
+ "learning_rate": 1.949424798228239e-08,
1824
+ "loss": 0.409,
1825
+ "step": 249
1826
+ },
1827
+ {
1828
+ "epoch": 2.4482338611449452,
1829
+ "grad_norm": 0.5454703569412231,
1830
+ "learning_rate": 4.873799534788059e-09,
1831
+ "loss": 0.3588,
1832
+ "step": 250
1833
+ },
1834
+ {
1835
+ "epoch": 2.4482338611449452,
1836
+ "eval_loss": 0.4610269367694855,
1837
+ "eval_runtime": 116.2672,
1838
+ "eval_samples_per_second": 7.371,
1839
+ "eval_steps_per_second": 0.464,
1840
+ "step": 250
1841
+ },
1842
+ {
1843
+ "epoch": 2.4482338611449452,
1844
+ "step": 250,
1845
+ "total_flos": 2.3792355349217935e+18,
1846
+ "train_loss": 0.49713707935810086,
1847
+ "train_runtime": 9531.3935,
1848
+ "train_samples_per_second": 2.518,
1849
+ "train_steps_per_second": 0.026
1850
+ }
1851
+ ],
1852
+ "logging_steps": 1,
1853
+ "max_steps": 250,
1854
+ "num_input_tokens_seen": 0,
1855
+ "num_train_epochs": 3,
1856
+ "save_steps": 50,
1857
+ "stateful_callbacks": {
1858
+ "TrainerControl": {
1859
+ "args": {
1860
+ "should_epoch_stop": false,
1861
+ "should_evaluate": false,
1862
+ "should_log": false,
1863
+ "should_save": true,
1864
+ "should_training_stop": true
1865
+ },
1866
+ "attributes": {}
1867
+ }
1868
+ },
1869
+ "total_flos": 2.3792355349217935e+18,
1870
+ "train_batch_size": 3,
1871
+ "trial_name": null,
1872
+ "trial_params": null
1873
+ }
TextUI-Func-7B/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d36469e2996ad0c3d0d1b2842ae5cdcc53ba9c32c2656bf00d763c2ed75ffd28
3
+ size 6161
TextUI-Func-7B/training_eval_loss.png ADDED
TextUI-Func-7B/training_loss.png ADDED
TextUI-Func-7B/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
TextUI-Trans-7B/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /root/autodl-fs/Qwen2.5-7B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.1
TextUI-Trans-7B/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/root/autodl-fs/Qwen2.5-7B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "down_proj",
28
+ "up_proj",
29
+ "o_proj",
30
+ "q_proj",
31
+ "k_proj",
32
+ "v_proj",
33
+ "gate_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
TextUI-Trans-7B/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:142327ece53e49d1a23f29308e93a0b5ab9c5c3f4e374ab7c941ce6c2640287d
3
+ size 161533192
TextUI-Trans-7B/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
TextUI-Trans-7B/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
TextUI-Trans-7B/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
TextUI-Trans-7B/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
TextUI-Trans-7B/tokenizer_config.json ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "right",
206
+ "split_special_tokens": false,
207
+ "tokenizer_class": "Qwen2Tokenizer",
208
+ "unk_token": null
209
+ }
TextUI-Trans-7B/trainer_log.jsonl ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"current_steps": 1, "total_steps": 300, "loss": 1.5618, "lr": 0.0, "epoch": 0.008537886872998933, "percentage": 0.33, "elapsed_time": "0:00:31", "remaining_time": "2:34:39"}
2
+ {"current_steps": 2, "total_steps": 300, "loss": 1.5366, "lr": 3.3333333333333333e-06, "epoch": 0.017075773745997867, "percentage": 0.67, "elapsed_time": "0:01:02", "remaining_time": "2:36:15"}
3
+ {"current_steps": 3, "total_steps": 300, "loss": 1.5274, "lr": 6.666666666666667e-06, "epoch": 0.025613660618996798, "percentage": 1.0, "elapsed_time": "0:01:31", "remaining_time": "2:31:07"}
4
+ {"current_steps": 4, "total_steps": 300, "loss": 1.3764, "lr": 1e-05, "epoch": 0.03415154749199573, "percentage": 1.33, "elapsed_time": "0:02:03", "remaining_time": "2:31:47"}
5
+ {"current_steps": 5, "total_steps": 300, "loss": 1.3908, "lr": 1.3333333333333333e-05, "epoch": 0.042689434364994665, "percentage": 1.67, "elapsed_time": "0:02:39", "remaining_time": "2:36:29"}
6
+ {"current_steps": 6, "total_steps": 300, "loss": 1.3127, "lr": 1.6666666666666667e-05, "epoch": 0.051227321237993596, "percentage": 2.0, "elapsed_time": "0:03:08", "remaining_time": "2:33:53"}
7
+ {"current_steps": 7, "total_steps": 300, "loss": 1.2693, "lr": 2e-05, "epoch": 0.05976520811099253, "percentage": 2.33, "elapsed_time": "0:03:43", "remaining_time": "2:35:58"}
8
+ {"current_steps": 8, "total_steps": 300, "loss": 1.3037, "lr": 2.3333333333333336e-05, "epoch": 0.06830309498399147, "percentage": 2.67, "elapsed_time": "0:04:09", "remaining_time": "2:31:40"}
9
+ {"current_steps": 9, "total_steps": 300, "loss": 1.1871, "lr": 2.6666666666666667e-05, "epoch": 0.0768409818569904, "percentage": 3.0, "elapsed_time": "0:04:35", "remaining_time": "2:28:32"}
10
+ {"current_steps": 10, "total_steps": 300, "loss": 1.15, "lr": 3e-05, "epoch": 0.08537886872998933, "percentage": 3.33, "elapsed_time": "0:05:06", "remaining_time": "2:27:58"}
11
+ {"current_steps": 11, "total_steps": 300, "loss": 1.2154, "lr": 3.3333333333333335e-05, "epoch": 0.09391675560298826, "percentage": 3.67, "elapsed_time": "0:05:40", "remaining_time": "2:29:08"}
12
+ {"current_steps": 12, "total_steps": 300, "loss": 1.1687, "lr": 3.6666666666666666e-05, "epoch": 0.10245464247598719, "percentage": 4.0, "elapsed_time": "0:06:08", "remaining_time": "2:27:16"}
13
+ {"current_steps": 13, "total_steps": 300, "loss": 1.2469, "lr": 4e-05, "epoch": 0.11099252934898612, "percentage": 4.33, "elapsed_time": "0:06:39", "remaining_time": "2:27:01"}
14
+ {"current_steps": 14, "total_steps": 300, "loss": 1.0827, "lr": 4.3333333333333334e-05, "epoch": 0.11953041622198506, "percentage": 4.67, "elapsed_time": "0:07:09", "remaining_time": "2:26:07"}
15
+ {"current_steps": 15, "total_steps": 300, "loss": 1.188, "lr": 4.666666666666667e-05, "epoch": 0.128068303094984, "percentage": 5.0, "elapsed_time": "0:07:42", "remaining_time": "2:26:35"}
16
+ {"current_steps": 16, "total_steps": 300, "loss": 1.1076, "lr": 5e-05, "epoch": 0.13660618996798293, "percentage": 5.33, "elapsed_time": "0:08:23", "remaining_time": "2:28:50"}
17
+ {"current_steps": 17, "total_steps": 300, "loss": 1.047, "lr": 5.333333333333333e-05, "epoch": 0.14514407684098185, "percentage": 5.67, "elapsed_time": "0:09:01", "remaining_time": "2:30:18"}
18
+ {"current_steps": 18, "total_steps": 300, "loss": 1.0498, "lr": 5.666666666666667e-05, "epoch": 0.1536819637139808, "percentage": 6.0, "elapsed_time": "0:09:43", "remaining_time": "2:32:19"}
19
+ {"current_steps": 19, "total_steps": 300, "loss": 1.0226, "lr": 6e-05, "epoch": 0.1622198505869797, "percentage": 6.33, "elapsed_time": "0:10:10", "remaining_time": "2:30:23"}
20
+ {"current_steps": 20, "total_steps": 300, "loss": 1.0192, "lr": 6.333333333333333e-05, "epoch": 0.17075773745997866, "percentage": 6.67, "elapsed_time": "0:10:40", "remaining_time": "2:29:23"}
21
+ {"current_steps": 21, "total_steps": 300, "loss": 1.0255, "lr": 6.666666666666667e-05, "epoch": 0.17929562433297758, "percentage": 7.0, "elapsed_time": "0:11:20", "remaining_time": "2:30:39"}
22
+ {"current_steps": 22, "total_steps": 300, "loss": 1.0127, "lr": 7e-05, "epoch": 0.18783351120597652, "percentage": 7.33, "elapsed_time": "0:11:47", "remaining_time": "2:29:04"}
23
+ {"current_steps": 23, "total_steps": 300, "loss": 1.0648, "lr": 7.333333333333333e-05, "epoch": 0.19637139807897544, "percentage": 7.67, "elapsed_time": "0:12:18", "remaining_time": "2:28:11"}
24
+ {"current_steps": 24, "total_steps": 300, "loss": 0.9899, "lr": 7.666666666666667e-05, "epoch": 0.20490928495197439, "percentage": 8.0, "elapsed_time": "0:12:49", "remaining_time": "2:27:34"}
25
+ {"current_steps": 25, "total_steps": 300, "loss": 0.9888, "lr": 8e-05, "epoch": 0.21344717182497333, "percentage": 8.33, "elapsed_time": "0:13:24", "remaining_time": "2:27:26"}
26
+ {"current_steps": 25, "total_steps": 300, "eval_loss": 0.9922655820846558, "epoch": 0.21344717182497333, "percentage": 8.33, "elapsed_time": "0:15:30", "remaining_time": "2:50:36"}
27
+ {"current_steps": 26, "total_steps": 300, "loss": 0.9876, "lr": 8.333333333333334e-05, "epoch": 0.22198505869797225, "percentage": 8.67, "elapsed_time": "0:16:06", "remaining_time": "2:49:44"}
28
+ {"current_steps": 27, "total_steps": 300, "loss": 0.9778, "lr": 8.666666666666667e-05, "epoch": 0.2305229455709712, "percentage": 9.0, "elapsed_time": "0:16:41", "remaining_time": "2:48:44"}
29
+ {"current_steps": 28, "total_steps": 300, "loss": 0.9561, "lr": 9e-05, "epoch": 0.2390608324439701, "percentage": 9.33, "elapsed_time": "0:17:13", "remaining_time": "2:47:19"}
30
+ {"current_steps": 29, "total_steps": 300, "loss": 0.9958, "lr": 9.333333333333334e-05, "epoch": 0.24759871931696906, "percentage": 9.67, "elapsed_time": "0:17:40", "remaining_time": "2:45:08"}
31
+ {"current_steps": 30, "total_steps": 300, "loss": 1.0282, "lr": 9.666666666666667e-05, "epoch": 0.256136606189968, "percentage": 10.0, "elapsed_time": "0:18:13", "remaining_time": "2:44:03"}
32
+ {"current_steps": 31, "total_steps": 300, "loss": 0.9626, "lr": 0.0001, "epoch": 0.2646744930629669, "percentage": 10.33, "elapsed_time": "0:18:40", "remaining_time": "2:42:02"}
33
+ {"current_steps": 32, "total_steps": 300, "loss": 1.0061, "lr": 9.999661540018812e-05, "epoch": 0.27321237993596587, "percentage": 10.67, "elapsed_time": "0:19:05", "remaining_time": "2:39:50"}
34
+ {"current_steps": 33, "total_steps": 300, "loss": 0.9061, "lr": 9.998646205897309e-05, "epoch": 0.28175026680896476, "percentage": 11.0, "elapsed_time": "0:19:56", "remaining_time": "2:41:24"}
35
+ {"current_steps": 34, "total_steps": 300, "loss": 0.9743, "lr": 9.99695413509548e-05, "epoch": 0.2902881536819637, "percentage": 11.33, "elapsed_time": "0:20:27", "remaining_time": "2:40:06"}
36
+ {"current_steps": 35, "total_steps": 300, "loss": 0.9536, "lr": 9.994585556692624e-05, "epoch": 0.29882604055496265, "percentage": 11.67, "elapsed_time": "0:21:02", "remaining_time": "2:39:18"}
37
+ {"current_steps": 36, "total_steps": 300, "loss": 0.998, "lr": 9.991540791356342e-05, "epoch": 0.3073639274279616, "percentage": 12.0, "elapsed_time": "0:21:33", "remaining_time": "2:38:08"}
38
+ {"current_steps": 37, "total_steps": 300, "loss": 1.0049, "lr": 9.987820251299122e-05, "epoch": 0.31590181430096054, "percentage": 12.33, "elapsed_time": "0:22:00", "remaining_time": "2:36:26"}
39
+ {"current_steps": 38, "total_steps": 300, "loss": 0.9955, "lr": 9.983424440222531e-05, "epoch": 0.3244397011739594, "percentage": 12.67, "elapsed_time": "0:22:30", "remaining_time": "2:35:13"}
40
+ {"current_steps": 39, "total_steps": 300, "loss": 0.9684, "lr": 9.978353953249022e-05, "epoch": 0.3329775880469584, "percentage": 13.0, "elapsed_time": "0:23:00", "remaining_time": "2:34:00"}
41
+ {"current_steps": 40, "total_steps": 300, "loss": 1.0183, "lr": 9.972609476841367e-05, "epoch": 0.3415154749199573, "percentage": 13.33, "elapsed_time": "0:23:32", "remaining_time": "2:33:01"}
42
+ {"current_steps": 41, "total_steps": 300, "loss": 0.9516, "lr": 9.966191788709716e-05, "epoch": 0.35005336179295626, "percentage": 13.67, "elapsed_time": "0:24:01", "remaining_time": "2:31:46"}
43
+ {"current_steps": 42, "total_steps": 300, "loss": 0.9607, "lr": 9.959101757706308e-05, "epoch": 0.35859124866595515, "percentage": 14.0, "elapsed_time": "0:24:32", "remaining_time": "2:30:46"}
44
+ {"current_steps": 43, "total_steps": 300, "loss": 0.9116, "lr": 9.951340343707852e-05, "epoch": 0.3671291355389541, "percentage": 14.33, "elapsed_time": "0:25:04", "remaining_time": "2:29:51"}
45
+ {"current_steps": 44, "total_steps": 300, "loss": 1.0001, "lr": 9.942908597485558e-05, "epoch": 0.37566702241195304, "percentage": 14.67, "elapsed_time": "0:25:39", "remaining_time": "2:29:14"}
46
+ {"current_steps": 45, "total_steps": 300, "loss": 0.9237, "lr": 9.933807660562898e-05, "epoch": 0.384204909284952, "percentage": 15.0, "elapsed_time": "0:26:10", "remaining_time": "2:28:21"}
47
+ {"current_steps": 46, "total_steps": 300, "loss": 0.9682, "lr": 9.924038765061042e-05, "epoch": 0.3927427961579509, "percentage": 15.33, "elapsed_time": "0:26:45", "remaining_time": "2:27:46"}
48
+ {"current_steps": 47, "total_steps": 300, "loss": 0.9255, "lr": 9.913603233532067e-05, "epoch": 0.4012806830309498, "percentage": 15.67, "elapsed_time": "0:27:15", "remaining_time": "2:26:45"}
49
+ {"current_steps": 48, "total_steps": 300, "loss": 0.8936, "lr": 9.902502478779896e-05, "epoch": 0.40981856990394877, "percentage": 16.0, "elapsed_time": "0:27:45", "remaining_time": "2:25:42"}
50
+ {"current_steps": 49, "total_steps": 300, "loss": 0.953, "lr": 9.890738003669029e-05, "epoch": 0.4183564567769477, "percentage": 16.33, "elapsed_time": "0:28:20", "remaining_time": "2:25:08"}
51
+ {"current_steps": 50, "total_steps": 300, "loss": 0.9401, "lr": 9.878311400921072e-05, "epoch": 0.42689434364994666, "percentage": 16.67, "elapsed_time": "0:28:48", "remaining_time": "2:24:02"}
52
+ {"current_steps": 50, "total_steps": 300, "eval_loss": 0.9200984835624695, "epoch": 0.42689434364994666, "percentage": 16.67, "elapsed_time": "0:30:54", "remaining_time": "2:34:34"}
53
+ {"current_steps": 51, "total_steps": 300, "loss": 0.8947, "lr": 9.865224352899119e-05, "epoch": 0.43543223052294555, "percentage": 17.0, "elapsed_time": "0:31:24", "remaining_time": "2:33:21"}
54
+ {"current_steps": 52, "total_steps": 300, "loss": 0.9038, "lr": 9.851478631379982e-05, "epoch": 0.4439701173959445, "percentage": 17.33, "elapsed_time": "0:31:58", "remaining_time": "2:32:30"}
55
+ {"current_steps": 53, "total_steps": 300, "loss": 0.9348, "lr": 9.837076097314319e-05, "epoch": 0.45250800426894344, "percentage": 17.67, "elapsed_time": "0:32:31", "remaining_time": "2:31:32"}
56
+ {"current_steps": 54, "total_steps": 300, "loss": 0.9275, "lr": 9.822018700574695e-05, "epoch": 0.4610458911419424, "percentage": 18.0, "elapsed_time": "0:33:07", "remaining_time": "2:30:52"}
57
+ {"current_steps": 55, "total_steps": 300, "loss": 0.928, "lr": 9.806308479691595e-05, "epoch": 0.4695837780149413, "percentage": 18.33, "elapsed_time": "0:33:42", "remaining_time": "2:30:07"}
58
+ {"current_steps": 56, "total_steps": 300, "loss": 0.9702, "lr": 9.789947561577445e-05, "epoch": 0.4781216648879402, "percentage": 18.67, "elapsed_time": "0:34:14", "remaining_time": "2:29:13"}
59
+ {"current_steps": 57, "total_steps": 300, "loss": 0.9534, "lr": 9.77293816123866e-05, "epoch": 0.48665955176093917, "percentage": 19.0, "elapsed_time": "0:34:51", "remaining_time": "2:28:37"}
60
+ {"current_steps": 58, "total_steps": 300, "loss": 0.9474, "lr": 9.755282581475769e-05, "epoch": 0.4951974386339381, "percentage": 19.33, "elapsed_time": "0:35:25", "remaining_time": "2:27:50"}
61
+ {"current_steps": 59, "total_steps": 300, "loss": 0.8856, "lr": 9.736983212571646e-05, "epoch": 0.503735325506937, "percentage": 19.67, "elapsed_time": "0:35:54", "remaining_time": "2:26:40"}
62
+ {"current_steps": 60, "total_steps": 300, "loss": 0.946, "lr": 9.718042531967918e-05, "epoch": 0.512273212379936, "percentage": 20.0, "elapsed_time": "0:36:22", "remaining_time": "2:25:29"}
63
+ {"current_steps": 61, "total_steps": 300, "loss": 0.9204, "lr": 9.698463103929542e-05, "epoch": 0.5208110992529349, "percentage": 20.33, "elapsed_time": "0:36:56", "remaining_time": "2:24:42"}
64
+ {"current_steps": 62, "total_steps": 300, "loss": 0.9127, "lr": 9.678247579197657e-05, "epoch": 0.5293489861259338, "percentage": 20.67, "elapsed_time": "0:37:26", "remaining_time": "2:23:45"}
65
+ {"current_steps": 63, "total_steps": 300, "loss": 0.8976, "lr": 9.657398694630712e-05, "epoch": 0.5378868729989328, "percentage": 21.0, "elapsed_time": "0:37:58", "remaining_time": "2:22:51"}
66
+ {"current_steps": 64, "total_steps": 300, "loss": 0.8394, "lr": 9.635919272833938e-05, "epoch": 0.5464247598719317, "percentage": 21.33, "elapsed_time": "0:38:32", "remaining_time": "2:22:08"}
67
+ {"current_steps": 65, "total_steps": 300, "loss": 0.9153, "lr": 9.613812221777212e-05, "epoch": 0.5549626467449307, "percentage": 21.67, "elapsed_time": "0:39:00", "remaining_time": "2:21:02"}
68
+ {"current_steps": 66, "total_steps": 300, "loss": 0.9224, "lr": 9.591080534401371e-05, "epoch": 0.5635005336179295, "percentage": 22.0, "elapsed_time": "0:39:27", "remaining_time": "2:19:53"}
69
+ {"current_steps": 67, "total_steps": 300, "loss": 0.9281, "lr": 9.567727288213005e-05, "epoch": 0.5720384204909285, "percentage": 22.33, "elapsed_time": "0:40:02", "remaining_time": "2:19:13"}
70
+ {"current_steps": 68, "total_steps": 300, "loss": 0.8459, "lr": 9.543755644867822e-05, "epoch": 0.5805763073639274, "percentage": 22.67, "elapsed_time": "0:40:28", "remaining_time": "2:18:04"}
71
+ {"current_steps": 69, "total_steps": 300, "loss": 0.9076, "lr": 9.519168849742604e-05, "epoch": 0.5891141942369263, "percentage": 23.0, "elapsed_time": "0:41:02", "remaining_time": "2:17:23"}
72
+ {"current_steps": 70, "total_steps": 300, "loss": 0.9292, "lr": 9.493970231495835e-05, "epoch": 0.5976520811099253, "percentage": 23.33, "elapsed_time": "0:41:36", "remaining_time": "2:16:41"}
73
+ {"current_steps": 71, "total_steps": 300, "loss": 0.7814, "lr": 9.468163201617062e-05, "epoch": 0.6061899679829242, "percentage": 23.67, "elapsed_time": "0:42:11", "remaining_time": "2:16:03"}
74
+ {"current_steps": 72, "total_steps": 300, "loss": 0.8878, "lr": 9.441751253965021e-05, "epoch": 0.6147278548559232, "percentage": 24.0, "elapsed_time": "0:42:45", "remaining_time": "2:15:24"}
75
+ {"current_steps": 73, "total_steps": 300, "loss": 0.9064, "lr": 9.414737964294636e-05, "epoch": 0.6232657417289221, "percentage": 24.33, "elapsed_time": "0:43:22", "remaining_time": "2:14:52"}
76
+ {"current_steps": 74, "total_steps": 300, "loss": 0.9185, "lr": 9.38712698977291e-05, "epoch": 0.6318036286019211, "percentage": 24.67, "elapsed_time": "0:43:56", "remaining_time": "2:14:10"}
77
+ {"current_steps": 75, "total_steps": 300, "loss": 0.9435, "lr": 9.358922068483812e-05, "epoch": 0.6403415154749199, "percentage": 25.0, "elapsed_time": "0:44:25", "remaining_time": "2:13:17"}
78
+ {"current_steps": 75, "total_steps": 300, "eval_loss": 0.8910433650016785, "epoch": 0.6403415154749199, "percentage": 25.0, "elapsed_time": "0:46:32", "remaining_time": "2:19:36"}
79
+ {"current_steps": 76, "total_steps": 300, "loss": 0.91, "lr": 9.330127018922194e-05, "epoch": 0.6488794023479189, "percentage": 25.33, "elapsed_time": "0:47:09", "remaining_time": "2:18:58"}
80
+ {"current_steps": 77, "total_steps": 300, "loss": 0.9656, "lr": 9.300745739476829e-05, "epoch": 0.6574172892209178, "percentage": 25.67, "elapsed_time": "0:47:39", "remaining_time": "2:18:02"}
81
+ {"current_steps": 78, "total_steps": 300, "loss": 0.9734, "lr": 9.270782207902629e-05, "epoch": 0.6659551760939167, "percentage": 26.0, "elapsed_time": "0:48:11", "remaining_time": "2:17:09"}
82
+ {"current_steps": 79, "total_steps": 300, "loss": 0.9102, "lr": 9.24024048078213e-05, "epoch": 0.6744930629669157, "percentage": 26.33, "elapsed_time": "0:48:48", "remaining_time": "2:16:33"}
83
+ {"current_steps": 80, "total_steps": 300, "loss": 0.9166, "lr": 9.209124692976287e-05, "epoch": 0.6830309498399146, "percentage": 26.67, "elapsed_time": "0:49:21", "remaining_time": "2:15:43"}
84
+ {"current_steps": 81, "total_steps": 300, "loss": 0.8672, "lr": 9.177439057064683e-05, "epoch": 0.6915688367129136, "percentage": 27.0, "elapsed_time": "0:49:53", "remaining_time": "2:14:53"}
85
+ {"current_steps": 82, "total_steps": 300, "loss": 0.8953, "lr": 9.145187862775209e-05, "epoch": 0.7001067235859125, "percentage": 27.33, "elapsed_time": "0:50:24", "remaining_time": "2:14:01"}
86
+ {"current_steps": 83, "total_steps": 300, "loss": 0.914, "lr": 9.112375476403312e-05, "epoch": 0.7086446104589115, "percentage": 27.67, "elapsed_time": "0:50:52", "remaining_time": "2:13:01"}
87
+ {"current_steps": 84, "total_steps": 300, "loss": 0.8936, "lr": 9.079006340220862e-05, "epoch": 0.7171824973319103, "percentage": 28.0, "elapsed_time": "0:51:29", "remaining_time": "2:12:23"}
88
+ {"current_steps": 85, "total_steps": 300, "loss": 0.9096, "lr": 9.045084971874738e-05, "epoch": 0.7257203842049093, "percentage": 28.33, "elapsed_time": "0:52:07", "remaining_time": "2:11:49"}
89
+ {"current_steps": 86, "total_steps": 300, "loss": 0.957, "lr": 9.01061596377522e-05, "epoch": 0.7342582710779082, "percentage": 28.67, "elapsed_time": "0:52:35", "remaining_time": "2:10:52"}
90
+ {"current_steps": 87, "total_steps": 300, "loss": 0.8853, "lr": 8.97560398247424e-05, "epoch": 0.7427961579509071, "percentage": 29.0, "elapsed_time": "0:53:07", "remaining_time": "2:10:04"}
91
+ {"current_steps": 88, "total_steps": 300, "loss": 0.907, "lr": 8.940053768033609e-05, "epoch": 0.7513340448239061, "percentage": 29.33, "elapsed_time": "0:53:38", "remaining_time": "2:09:13"}
92
+ {"current_steps": 89, "total_steps": 300, "loss": 0.8719, "lr": 8.903970133383297e-05, "epoch": 0.759871931696905, "percentage": 29.67, "elapsed_time": "0:54:10", "remaining_time": "2:08:25"}
93
+ {"current_steps": 90, "total_steps": 300, "loss": 0.9782, "lr": 8.86735796366982e-05, "epoch": 0.768409818569904, "percentage": 30.0, "elapsed_time": "0:54:44", "remaining_time": "2:07:43"}
94
+ {"current_steps": 91, "total_steps": 300, "loss": 0.9225, "lr": 8.83022221559489e-05, "epoch": 0.7769477054429029, "percentage": 30.33, "elapsed_time": "0:55:17", "remaining_time": "2:06:58"}
95
+ {"current_steps": 92, "total_steps": 300, "loss": 0.9045, "lr": 8.792567916744346e-05, "epoch": 0.7854855923159018, "percentage": 30.67, "elapsed_time": "0:55:52", "remaining_time": "2:06:19"}
96
+ {"current_steps": 93, "total_steps": 300, "loss": 0.9335, "lr": 8.754400164907497e-05, "epoch": 0.7940234791889007, "percentage": 31.0, "elapsed_time": "0:56:31", "remaining_time": "2:05:48"}
97
+ {"current_steps": 94, "total_steps": 300, "loss": 0.9086, "lr": 8.715724127386972e-05, "epoch": 0.8025613660618997, "percentage": 31.33, "elapsed_time": "0:56:56", "remaining_time": "2:04:47"}
98
+ {"current_steps": 95, "total_steps": 300, "loss": 0.9079, "lr": 8.676545040299145e-05, "epoch": 0.8110992529348986, "percentage": 31.67, "elapsed_time": "0:57:23", "remaining_time": "2:03:51"}
99
+ {"current_steps": 96, "total_steps": 300, "loss": 0.9059, "lr": 8.636868207865244e-05, "epoch": 0.8196371398078975, "percentage": 32.0, "elapsed_time": "0:57:53", "remaining_time": "2:03:00"}
100
+ {"current_steps": 97, "total_steps": 300, "loss": 0.847, "lr": 8.596699001693255e-05, "epoch": 0.8281750266808965, "percentage": 32.33, "elapsed_time": "0:58:24", "remaining_time": "2:02:15"}
101
+ {"current_steps": 98, "total_steps": 300, "loss": 0.911, "lr": 8.556042860050687e-05, "epoch": 0.8367129135538954, "percentage": 32.67, "elapsed_time": "0:58:55", "remaining_time": "2:01:28"}
102
+ {"current_steps": 99, "total_steps": 300, "loss": 0.8864, "lr": 8.51490528712831e-05, "epoch": 0.8452508004268944, "percentage": 33.0, "elapsed_time": "0:59:26", "remaining_time": "2:00:41"}
103
+ {"current_steps": 100, "total_steps": 300, "loss": 0.8479, "lr": 8.473291852294987e-05, "epoch": 0.8537886872998933, "percentage": 33.33, "elapsed_time": "0:59:56", "remaining_time": "1:59:52"}
104
+ {"current_steps": 100, "total_steps": 300, "eval_loss": 0.8699346780776978, "epoch": 0.8537886872998933, "percentage": 33.33, "elapsed_time": "1:02:02", "remaining_time": "2:04:04"}
105
+ {"current_steps": 101, "total_steps": 300, "loss": 0.92, "lr": 8.43120818934367e-05, "epoch": 0.8623265741728922, "percentage": 33.67, "elapsed_time": "1:02:30", "remaining_time": "2:03:10"}
106
+ {"current_steps": 102, "total_steps": 300, "loss": 0.8551, "lr": 8.388659995728663e-05, "epoch": 0.8708644610458911, "percentage": 34.0, "elapsed_time": "1:03:05", "remaining_time": "2:02:28"}
107
+ {"current_steps": 103, "total_steps": 300, "loss": 0.9024, "lr": 8.345653031794292e-05, "epoch": 0.87940234791889, "percentage": 34.33, "elapsed_time": "1:03:36", "remaining_time": "2:01:39"}
108
+ {"current_steps": 104, "total_steps": 300, "loss": 0.9081, "lr": 8.302193119995039e-05, "epoch": 0.887940234791889, "percentage": 34.67, "elapsed_time": "1:04:06", "remaining_time": "2:00:49"}
109
+ {"current_steps": 105, "total_steps": 300, "loss": 0.9215, "lr": 8.258286144107276e-05, "epoch": 0.8964781216648879, "percentage": 35.0, "elapsed_time": "1:04:34", "remaining_time": "1:59:55"}
110
+ {"current_steps": 106, "total_steps": 300, "loss": 0.886, "lr": 8.213938048432697e-05, "epoch": 0.9050160085378869, "percentage": 35.33, "elapsed_time": "1:05:05", "remaining_time": "1:59:08"}
111
+ {"current_steps": 107, "total_steps": 300, "loss": 0.8436, "lr": 8.169154836993551e-05, "epoch": 0.9135538954108858, "percentage": 35.67, "elapsed_time": "1:05:35", "remaining_time": "1:58:18"}
112
+ {"current_steps": 108, "total_steps": 300, "loss": 0.8181, "lr": 8.1239425727198e-05, "epoch": 0.9220917822838848, "percentage": 36.0, "elapsed_time": "1:06:05", "remaining_time": "1:57:29"}
113
+ {"current_steps": 109, "total_steps": 300, "loss": 0.8687, "lr": 8.07830737662829e-05, "epoch": 0.9306296691568837, "percentage": 36.33, "elapsed_time": "1:06:29", "remaining_time": "1:56:31"}
114
+ {"current_steps": 110, "total_steps": 300, "loss": 0.8341, "lr": 8.032255426994069e-05, "epoch": 0.9391675560298826, "percentage": 36.67, "elapsed_time": "1:06:58", "remaining_time": "1:55:40"}
115
+ {"current_steps": 111, "total_steps": 300, "loss": 0.9307, "lr": 7.985792958513931e-05, "epoch": 0.9477054429028815, "percentage": 37.0, "elapsed_time": "1:07:26", "remaining_time": "1:54:49"}
116
+ {"current_steps": 112, "total_steps": 300, "loss": 0.8765, "lr": 7.938926261462366e-05, "epoch": 0.9562433297758804, "percentage": 37.33, "elapsed_time": "1:07:56", "remaining_time": "1:54:03"}
117
+ {"current_steps": 113, "total_steps": 300, "loss": 0.8956, "lr": 7.891661680839932e-05, "epoch": 0.9647812166488794, "percentage": 37.67, "elapsed_time": "1:08:32", "remaining_time": "1:53:25"}
118
+ {"current_steps": 114, "total_steps": 300, "loss": 0.8607, "lr": 7.844005615514259e-05, "epoch": 0.9733191035218783, "percentage": 38.0, "elapsed_time": "1:09:00", "remaining_time": "1:52:35"}
119
+ {"current_steps": 115, "total_steps": 300, "loss": 0.8357, "lr": 7.795964517353735e-05, "epoch": 0.9818569903948773, "percentage": 38.33, "elapsed_time": "1:09:31", "remaining_time": "1:51:51"}
120
+ {"current_steps": 116, "total_steps": 300, "loss": 0.9093, "lr": 7.74754489035403e-05, "epoch": 0.9903948772678762, "percentage": 38.67, "elapsed_time": "1:10:06", "remaining_time": "1:51:11"}
121
+ {"current_steps": 117, "total_steps": 300, "loss": 0.8595, "lr": 7.698753289757565e-05, "epoch": 0.9989327641408752, "percentage": 39.0, "elapsed_time": "1:10:36", "remaining_time": "1:50:26"}
122
+ {"current_steps": 118, "total_steps": 300, "loss": 0.968, "lr": 7.649596321166024e-05, "epoch": 1.0, "percentage": 39.33, "elapsed_time": "1:10:40", "remaining_time": "1:49:01"}
123
+ {"current_steps": 119, "total_steps": 300, "loss": 0.8389, "lr": 7.600080639646077e-05, "epoch": 1.0085378868729988, "percentage": 39.67, "elapsed_time": "1:11:11", "remaining_time": "1:48:16"}
124
+ {"current_steps": 120, "total_steps": 300, "loss": 0.8826, "lr": 7.550212948828377e-05, "epoch": 1.017075773745998, "percentage": 40.0, "elapsed_time": "1:11:46", "remaining_time": "1:47:39"}
125
+ {"current_steps": 121, "total_steps": 300, "loss": 0.8961, "lr": 7.500000000000001e-05, "epoch": 1.0256136606189967, "percentage": 40.33, "elapsed_time": "1:12:13", "remaining_time": "1:46:51"}
126
+ {"current_steps": 122, "total_steps": 300, "loss": 0.7779, "lr": 7.449448591190435e-05, "epoch": 1.0341515474919958, "percentage": 40.67, "elapsed_time": "1:12:46", "remaining_time": "1:46:10"}
127
+ {"current_steps": 123, "total_steps": 300, "loss": 0.882, "lr": 7.398565566251232e-05, "epoch": 1.0426894343649946, "percentage": 41.0, "elapsed_time": "1:13:15", "remaining_time": "1:45:25"}
128
+ {"current_steps": 124, "total_steps": 300, "loss": 0.8583, "lr": 7.347357813929454e-05, "epoch": 1.0512273212379937, "percentage": 41.33, "elapsed_time": "1:13:46", "remaining_time": "1:44:43"}
129
+ {"current_steps": 125, "total_steps": 300, "loss": 0.8085, "lr": 7.295832266935059e-05, "epoch": 1.0597652081109925, "percentage": 41.67, "elapsed_time": "1:14:15", "remaining_time": "1:43:57"}
130
+ {"current_steps": 125, "total_steps": 300, "eval_loss": 0.855903685092926, "epoch": 1.0597652081109925, "percentage": 41.67, "elapsed_time": "1:16:21", "remaining_time": "1:46:54"}
131
+ {"current_steps": 126, "total_steps": 300, "loss": 0.8287, "lr": 7.243995901002312e-05, "epoch": 1.0683030949839916, "percentage": 42.0, "elapsed_time": "1:16:51", "remaining_time": "1:46:08"}
132
+ {"current_steps": 127, "total_steps": 300, "loss": 0.8779, "lr": 7.191855733945387e-05, "epoch": 1.0768409818569904, "percentage": 42.33, "elapsed_time": "1:17:20", "remaining_time": "1:45:20"}
133
+ {"current_steps": 128, "total_steps": 300, "loss": 0.8207, "lr": 7.139418824708272e-05, "epoch": 1.0853788687299892, "percentage": 42.67, "elapsed_time": "1:17:55", "remaining_time": "1:44:43"}
134
+ {"current_steps": 129, "total_steps": 300, "loss": 0.8071, "lr": 7.08669227240909e-05, "epoch": 1.0939167556029883, "percentage": 43.0, "elapsed_time": "1:18:30", "remaining_time": "1:44:04"}
135
+ {"current_steps": 130, "total_steps": 300, "loss": 0.816, "lr": 7.033683215379002e-05, "epoch": 1.1024546424759871, "percentage": 43.33, "elapsed_time": "1:19:04", "remaining_time": "1:43:23"}
136
+ {"current_steps": 131, "total_steps": 300, "loss": 0.8307, "lr": 6.980398830195785e-05, "epoch": 1.1109925293489862, "percentage": 43.67, "elapsed_time": "1:19:33", "remaining_time": "1:42:38"}
137
+ {"current_steps": 132, "total_steps": 300, "loss": 0.8799, "lr": 6.926846330712242e-05, "epoch": 1.119530416221985, "percentage": 44.0, "elapsed_time": "1:20:03", "remaining_time": "1:41:53"}
138
+ {"current_steps": 133, "total_steps": 300, "loss": 0.8427, "lr": 6.873032967079561e-05, "epoch": 1.128068303094984, "percentage": 44.33, "elapsed_time": "1:20:28", "remaining_time": "1:41:03"}
139
+ {"current_steps": 134, "total_steps": 300, "loss": 0.857, "lr": 6.818966024765758e-05, "epoch": 1.136606189967983, "percentage": 44.67, "elapsed_time": "1:21:00", "remaining_time": "1:40:20"}
140
+ {"current_steps": 135, "total_steps": 300, "loss": 0.8238, "lr": 6.764652823569344e-05, "epoch": 1.1451440768409817, "percentage": 45.0, "elapsed_time": "1:21:37", "remaining_time": "1:39:45"}
141
+ {"current_steps": 136, "total_steps": 300, "loss": 0.8208, "lr": 6.710100716628344e-05, "epoch": 1.1536819637139808, "percentage": 45.33, "elapsed_time": "1:22:11", "remaining_time": "1:39:06"}
142
+ {"current_steps": 137, "total_steps": 300, "loss": 0.7936, "lr": 6.65531708942479e-05, "epoch": 1.1622198505869796, "percentage": 45.67, "elapsed_time": "1:22:40", "remaining_time": "1:38:22"}
143
+ {"current_steps": 138, "total_steps": 300, "loss": 0.8355, "lr": 6.600309358784857e-05, "epoch": 1.1707577374599787, "percentage": 46.0, "elapsed_time": "1:23:20", "remaining_time": "1:37:49"}
144
+ {"current_steps": 139, "total_steps": 300, "loss": 0.7518, "lr": 6.545084971874738e-05, "epoch": 1.1792956243329775, "percentage": 46.33, "elapsed_time": "1:23:47", "remaining_time": "1:37:03"}
145
+ {"current_steps": 140, "total_steps": 300, "loss": 0.8091, "lr": 6.48965140519241e-05, "epoch": 1.1878335112059766, "percentage": 46.67, "elapsed_time": "1:24:23", "remaining_time": "1:36:26"}
146
+ {"current_steps": 141, "total_steps": 300, "loss": 0.8138, "lr": 6.434016163555452e-05, "epoch": 1.1963713980789754, "percentage": 47.0, "elapsed_time": "1:24:50", "remaining_time": "1:35:40"}
147
+ {"current_steps": 142, "total_steps": 300, "loss": 0.8357, "lr": 6.378186779084995e-05, "epoch": 1.2049092849519745, "percentage": 47.33, "elapsed_time": "1:25:26", "remaining_time": "1:35:03"}
148
+ {"current_steps": 143, "total_steps": 300, "loss": 0.8278, "lr": 6.322170810186012e-05, "epoch": 1.2134471718249733, "percentage": 47.67, "elapsed_time": "1:25:52", "remaining_time": "1:34:17"}
149
+ {"current_steps": 144, "total_steps": 300, "loss": 0.8395, "lr": 6.26597584052401e-05, "epoch": 1.2219850586979724, "percentage": 48.0, "elapsed_time": "1:26:19", "remaining_time": "1:33:30"}
150
+ {"current_steps": 145, "total_steps": 300, "loss": 0.811, "lr": 6.209609477998338e-05, "epoch": 1.2305229455709712, "percentage": 48.33, "elapsed_time": "1:26:53", "remaining_time": "1:32:52"}
151
+ {"current_steps": 146, "total_steps": 300, "loss": 0.8305, "lr": 6.153079353712201e-05, "epoch": 1.23906083244397, "percentage": 48.67, "elapsed_time": "1:27:19", "remaining_time": "1:32:06"}
152
+ {"current_steps": 147, "total_steps": 300, "loss": 0.8235, "lr": 6.096393120939516e-05, "epoch": 1.247598719316969, "percentage": 49.0, "elapsed_time": "1:27:48", "remaining_time": "1:31:23"}
153
+ {"current_steps": 148, "total_steps": 300, "loss": 0.8181, "lr": 6.0395584540887963e-05, "epoch": 1.256136606189968, "percentage": 49.33, "elapsed_time": "1:28:21", "remaining_time": "1:30:44"}
154
+ {"current_steps": 149, "total_steps": 300, "loss": 0.8309, "lr": 5.982583047664151e-05, "epoch": 1.264674493062967, "percentage": 49.67, "elapsed_time": "1:28:55", "remaining_time": "1:30:07"}
155
+ {"current_steps": 150, "total_steps": 300, "loss": 0.8345, "lr": 5.925474615223573e-05, "epoch": 1.2732123799359658, "percentage": 50.0, "elapsed_time": "1:29:26", "remaining_time": "1:29:26"}
156
+ {"current_steps": 150, "total_steps": 300, "eval_loss": 0.8430656790733337, "epoch": 1.2732123799359658, "percentage": 50.0, "elapsed_time": "1:31:32", "remaining_time": "1:31:32"}
157
+ {"current_steps": 151, "total_steps": 300, "loss": 0.8162, "lr": 5.868240888334653e-05, "epoch": 1.2817502668089649, "percentage": 50.33, "elapsed_time": "1:32:03", "remaining_time": "1:30:50"}
158
+ {"current_steps": 152, "total_steps": 300, "loss": 0.8663, "lr": 5.810889615527838e-05, "epoch": 1.2902881536819637, "percentage": 50.67, "elapsed_time": "1:32:34", "remaining_time": "1:30:08"}
159
+ {"current_steps": 153, "total_steps": 300, "loss": 0.8306, "lr": 5.753428561247416e-05, "epoch": 1.2988260405549625, "percentage": 51.0, "elapsed_time": "1:33:02", "remaining_time": "1:29:23"}
160
+ {"current_steps": 154, "total_steps": 300, "loss": 0.7522, "lr": 5.695865504800327e-05, "epoch": 1.3073639274279616, "percentage": 51.33, "elapsed_time": "1:33:38", "remaining_time": "1:28:46"}
161
+ {"current_steps": 155, "total_steps": 300, "loss": 0.8092, "lr": 5.6382082393029746e-05, "epoch": 1.3159018143009606, "percentage": 51.67, "elapsed_time": "1:34:08", "remaining_time": "1:28:04"}
162
+ {"current_steps": 156, "total_steps": 300, "loss": 0.8326, "lr": 5.5804645706261514e-05, "epoch": 1.3244397011739595, "percentage": 52.0, "elapsed_time": "1:34:33", "remaining_time": "1:27:17"}
163
+ {"current_steps": 157, "total_steps": 300, "loss": 0.7984, "lr": 5.522642316338268e-05, "epoch": 1.3329775880469583, "percentage": 52.33, "elapsed_time": "1:35:03", "remaining_time": "1:26:34"}
164
+ {"current_steps": 158, "total_steps": 300, "loss": 0.8724, "lr": 5.464749304646962e-05, "epoch": 1.3415154749199574, "percentage": 52.67, "elapsed_time": "1:35:28", "remaining_time": "1:25:48"}
165
+ {"current_steps": 159, "total_steps": 300, "loss": 0.8529, "lr": 5.4067933733392915e-05, "epoch": 1.3500533617929562, "percentage": 53.0, "elapsed_time": "1:36:00", "remaining_time": "1:25:08"}
166
+ {"current_steps": 160, "total_steps": 300, "loss": 0.7884, "lr": 5.348782368720626e-05, "epoch": 1.358591248665955, "percentage": 53.33, "elapsed_time": "1:36:27", "remaining_time": "1:24:24"}
167
+ {"current_steps": 161, "total_steps": 300, "loss": 0.7394, "lr": 5.290724144552379e-05, "epoch": 1.367129135538954, "percentage": 53.67, "elapsed_time": "1:37:01", "remaining_time": "1:23:45"}
168
+ {"current_steps": 162, "total_steps": 300, "loss": 0.8742, "lr": 5.232626560988735e-05, "epoch": 1.3756670224119532, "percentage": 54.0, "elapsed_time": "1:37:39", "remaining_time": "1:23:11"}
169
+ {"current_steps": 163, "total_steps": 300, "loss": 0.8229, "lr": 5.174497483512506e-05, "epoch": 1.384204909284952, "percentage": 54.33, "elapsed_time": "1:38:17", "remaining_time": "1:22:36"}
TextUI-Trans-7B/trainer_state.json ADDED
@@ -0,0 +1,1132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 150,
3
+ "best_metric": 0.8430656790733337,
4
+ "best_model_checkpoint": "/root/autodl-tmp/model/lora-textui/stage2_cap_func_trans/checkpoint-150",
5
+ "epoch": 1.2732123799359658,
6
+ "eval_steps": 25,
7
+ "global_step": 150,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.008537886872998933,
14
+ "grad_norm": 3.569350242614746,
15
+ "learning_rate": 0.0,
16
+ "loss": 1.5618,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.017075773745997867,
21
+ "grad_norm": 3.5417115688323975,
22
+ "learning_rate": 3.3333333333333333e-06,
23
+ "loss": 1.5366,
24
+ "step": 2
25
+ },
26
+ {
27
+ "epoch": 0.025613660618996798,
28
+ "grad_norm": 3.087338924407959,
29
+ "learning_rate": 6.666666666666667e-06,
30
+ "loss": 1.5274,
31
+ "step": 3
32
+ },
33
+ {
34
+ "epoch": 0.03415154749199573,
35
+ "grad_norm": 2.873836040496826,
36
+ "learning_rate": 1e-05,
37
+ "loss": 1.3764,
38
+ "step": 4
39
+ },
40
+ {
41
+ "epoch": 0.042689434364994665,
42
+ "grad_norm": 2.4693570137023926,
43
+ "learning_rate": 1.3333333333333333e-05,
44
+ "loss": 1.3908,
45
+ "step": 5
46
+ },
47
+ {
48
+ "epoch": 0.051227321237993596,
49
+ "grad_norm": 1.9929084777832031,
50
+ "learning_rate": 1.6666666666666667e-05,
51
+ "loss": 1.3127,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 0.05976520811099253,
56
+ "grad_norm": 1.4752813577651978,
57
+ "learning_rate": 2e-05,
58
+ "loss": 1.2693,
59
+ "step": 7
60
+ },
61
+ {
62
+ "epoch": 0.06830309498399147,
63
+ "grad_norm": 1.0920029878616333,
64
+ "learning_rate": 2.3333333333333336e-05,
65
+ "loss": 1.3037,
66
+ "step": 8
67
+ },
68
+ {
69
+ "epoch": 0.0768409818569904,
70
+ "grad_norm": 0.8210710287094116,
71
+ "learning_rate": 2.6666666666666667e-05,
72
+ "loss": 1.1871,
73
+ "step": 9
74
+ },
75
+ {
76
+ "epoch": 0.08537886872998933,
77
+ "grad_norm": 0.7052881717681885,
78
+ "learning_rate": 3e-05,
79
+ "loss": 1.15,
80
+ "step": 10
81
+ },
82
+ {
83
+ "epoch": 0.09391675560298826,
84
+ "grad_norm": 0.6850194334983826,
85
+ "learning_rate": 3.3333333333333335e-05,
86
+ "loss": 1.2154,
87
+ "step": 11
88
+ },
89
+ {
90
+ "epoch": 0.10245464247598719,
91
+ "grad_norm": 0.6714601516723633,
92
+ "learning_rate": 3.6666666666666666e-05,
93
+ "loss": 1.1687,
94
+ "step": 12
95
+ },
96
+ {
97
+ "epoch": 0.11099252934898612,
98
+ "grad_norm": 0.6564128994941711,
99
+ "learning_rate": 4e-05,
100
+ "loss": 1.2469,
101
+ "step": 13
102
+ },
103
+ {
104
+ "epoch": 0.11953041622198506,
105
+ "grad_norm": 0.6847338080406189,
106
+ "learning_rate": 4.3333333333333334e-05,
107
+ "loss": 1.0827,
108
+ "step": 14
109
+ },
110
+ {
111
+ "epoch": 0.128068303094984,
112
+ "grad_norm": 0.5931199789047241,
113
+ "learning_rate": 4.666666666666667e-05,
114
+ "loss": 1.188,
115
+ "step": 15
116
+ },
117
+ {
118
+ "epoch": 0.13660618996798293,
119
+ "grad_norm": 0.5378791689872742,
120
+ "learning_rate": 5e-05,
121
+ "loss": 1.1076,
122
+ "step": 16
123
+ },
124
+ {
125
+ "epoch": 0.14514407684098185,
126
+ "grad_norm": 0.503239631652832,
127
+ "learning_rate": 5.333333333333333e-05,
128
+ "loss": 1.047,
129
+ "step": 17
130
+ },
131
+ {
132
+ "epoch": 0.1536819637139808,
133
+ "grad_norm": 0.4387917220592499,
134
+ "learning_rate": 5.666666666666667e-05,
135
+ "loss": 1.0498,
136
+ "step": 18
137
+ },
138
+ {
139
+ "epoch": 0.1622198505869797,
140
+ "grad_norm": 0.4416464865207672,
141
+ "learning_rate": 6e-05,
142
+ "loss": 1.0226,
143
+ "step": 19
144
+ },
145
+ {
146
+ "epoch": 0.17075773745997866,
147
+ "grad_norm": 0.44988617300987244,
148
+ "learning_rate": 6.333333333333333e-05,
149
+ "loss": 1.0192,
150
+ "step": 20
151
+ },
152
+ {
153
+ "epoch": 0.17929562433297758,
154
+ "grad_norm": 0.42736634612083435,
155
+ "learning_rate": 6.666666666666667e-05,
156
+ "loss": 1.0255,
157
+ "step": 21
158
+ },
159
+ {
160
+ "epoch": 0.18783351120597652,
161
+ "grad_norm": 0.43649032711982727,
162
+ "learning_rate": 7e-05,
163
+ "loss": 1.0127,
164
+ "step": 22
165
+ },
166
+ {
167
+ "epoch": 0.19637139807897544,
168
+ "grad_norm": 0.4321700632572174,
169
+ "learning_rate": 7.333333333333333e-05,
170
+ "loss": 1.0648,
171
+ "step": 23
172
+ },
173
+ {
174
+ "epoch": 0.20490928495197439,
175
+ "grad_norm": 0.4347881078720093,
176
+ "learning_rate": 7.666666666666667e-05,
177
+ "loss": 0.9899,
178
+ "step": 24
179
+ },
180
+ {
181
+ "epoch": 0.21344717182497333,
182
+ "grad_norm": 0.3858855962753296,
183
+ "learning_rate": 8e-05,
184
+ "loss": 0.9888,
185
+ "step": 25
186
+ },
187
+ {
188
+ "epoch": 0.21344717182497333,
189
+ "eval_loss": 0.9922655820846558,
190
+ "eval_runtime": 126.3706,
191
+ "eval_samples_per_second": 7.739,
192
+ "eval_steps_per_second": 0.491,
193
+ "step": 25
194
+ },
195
+ {
196
+ "epoch": 0.22198505869797225,
197
+ "grad_norm": 0.36364927887916565,
198
+ "learning_rate": 8.333333333333334e-05,
199
+ "loss": 0.9876,
200
+ "step": 26
201
+ },
202
+ {
203
+ "epoch": 0.2305229455709712,
204
+ "grad_norm": 0.3517715334892273,
205
+ "learning_rate": 8.666666666666667e-05,
206
+ "loss": 0.9778,
207
+ "step": 27
208
+ },
209
+ {
210
+ "epoch": 0.2390608324439701,
211
+ "grad_norm": 0.31786659359931946,
212
+ "learning_rate": 9e-05,
213
+ "loss": 0.9561,
214
+ "step": 28
215
+ },
216
+ {
217
+ "epoch": 0.24759871931696906,
218
+ "grad_norm": 0.34954020380973816,
219
+ "learning_rate": 9.333333333333334e-05,
220
+ "loss": 0.9958,
221
+ "step": 29
222
+ },
223
+ {
224
+ "epoch": 0.256136606189968,
225
+ "grad_norm": 0.3508739769458771,
226
+ "learning_rate": 9.666666666666667e-05,
227
+ "loss": 1.0282,
228
+ "step": 30
229
+ },
230
+ {
231
+ "epoch": 0.2646744930629669,
232
+ "grad_norm": 0.33114153146743774,
233
+ "learning_rate": 0.0001,
234
+ "loss": 0.9626,
235
+ "step": 31
236
+ },
237
+ {
238
+ "epoch": 0.27321237993596587,
239
+ "grad_norm": 0.3469662666320801,
240
+ "learning_rate": 9.999661540018812e-05,
241
+ "loss": 1.0061,
242
+ "step": 32
243
+ },
244
+ {
245
+ "epoch": 0.28175026680896476,
246
+ "grad_norm": 0.34702596068382263,
247
+ "learning_rate": 9.998646205897309e-05,
248
+ "loss": 0.9061,
249
+ "step": 33
250
+ },
251
+ {
252
+ "epoch": 0.2902881536819637,
253
+ "grad_norm": 0.30591893196105957,
254
+ "learning_rate": 9.99695413509548e-05,
255
+ "loss": 0.9743,
256
+ "step": 34
257
+ },
258
+ {
259
+ "epoch": 0.29882604055496265,
260
+ "grad_norm": 0.30477747321128845,
261
+ "learning_rate": 9.994585556692624e-05,
262
+ "loss": 0.9536,
263
+ "step": 35
264
+ },
265
+ {
266
+ "epoch": 0.3073639274279616,
267
+ "grad_norm": 0.31681594252586365,
268
+ "learning_rate": 9.991540791356342e-05,
269
+ "loss": 0.998,
270
+ "step": 36
271
+ },
272
+ {
273
+ "epoch": 0.31590181430096054,
274
+ "grad_norm": 0.30779823660850525,
275
+ "learning_rate": 9.987820251299122e-05,
276
+ "loss": 1.0049,
277
+ "step": 37
278
+ },
279
+ {
280
+ "epoch": 0.3244397011739594,
281
+ "grad_norm": 0.30715224146842957,
282
+ "learning_rate": 9.983424440222531e-05,
283
+ "loss": 0.9955,
284
+ "step": 38
285
+ },
286
+ {
287
+ "epoch": 0.3329775880469584,
288
+ "grad_norm": 0.30985817313194275,
289
+ "learning_rate": 9.978353953249022e-05,
290
+ "loss": 0.9684,
291
+ "step": 39
292
+ },
293
+ {
294
+ "epoch": 0.3415154749199573,
295
+ "grad_norm": 0.31904837489128113,
296
+ "learning_rate": 9.972609476841367e-05,
297
+ "loss": 1.0183,
298
+ "step": 40
299
+ },
300
+ {
301
+ "epoch": 0.35005336179295626,
302
+ "grad_norm": 0.29088112711906433,
303
+ "learning_rate": 9.966191788709716e-05,
304
+ "loss": 0.9516,
305
+ "step": 41
306
+ },
307
+ {
308
+ "epoch": 0.35859124866595515,
309
+ "grad_norm": 0.3119213283061981,
310
+ "learning_rate": 9.959101757706308e-05,
311
+ "loss": 0.9607,
312
+ "step": 42
313
+ },
314
+ {
315
+ "epoch": 0.3671291355389541,
316
+ "grad_norm": 0.2980335056781769,
317
+ "learning_rate": 9.951340343707852e-05,
318
+ "loss": 0.9116,
319
+ "step": 43
320
+ },
321
+ {
322
+ "epoch": 0.37566702241195304,
323
+ "grad_norm": 0.31713929772377014,
324
+ "learning_rate": 9.942908597485558e-05,
325
+ "loss": 1.0001,
326
+ "step": 44
327
+ },
328
+ {
329
+ "epoch": 0.384204909284952,
330
+ "grad_norm": 0.30609187483787537,
331
+ "learning_rate": 9.933807660562898e-05,
332
+ "loss": 0.9237,
333
+ "step": 45
334
+ },
335
+ {
336
+ "epoch": 0.3927427961579509,
337
+ "grad_norm": 0.30118709802627563,
338
+ "learning_rate": 9.924038765061042e-05,
339
+ "loss": 0.9682,
340
+ "step": 46
341
+ },
342
+ {
343
+ "epoch": 0.4012806830309498,
344
+ "grad_norm": 0.3028307855129242,
345
+ "learning_rate": 9.913603233532067e-05,
346
+ "loss": 0.9255,
347
+ "step": 47
348
+ },
349
+ {
350
+ "epoch": 0.40981856990394877,
351
+ "grad_norm": 0.31861209869384766,
352
+ "learning_rate": 9.902502478779896e-05,
353
+ "loss": 0.8936,
354
+ "step": 48
355
+ },
356
+ {
357
+ "epoch": 0.4183564567769477,
358
+ "grad_norm": 0.2734357416629791,
359
+ "learning_rate": 9.890738003669029e-05,
360
+ "loss": 0.953,
361
+ "step": 49
362
+ },
363
+ {
364
+ "epoch": 0.42689434364994666,
365
+ "grad_norm": 0.29771876335144043,
366
+ "learning_rate": 9.878311400921072e-05,
367
+ "loss": 0.9401,
368
+ "step": 50
369
+ },
370
+ {
371
+ "epoch": 0.42689434364994666,
372
+ "eval_loss": 0.9200984835624695,
373
+ "eval_runtime": 126.2659,
374
+ "eval_samples_per_second": 7.746,
375
+ "eval_steps_per_second": 0.491,
376
+ "step": 50
377
+ },
378
+ {
379
+ "epoch": 0.43543223052294555,
380
+ "grad_norm": 0.3155674338340759,
381
+ "learning_rate": 9.865224352899119e-05,
382
+ "loss": 0.8947,
383
+ "step": 51
384
+ },
385
+ {
386
+ "epoch": 0.4439701173959445,
387
+ "grad_norm": 0.30623170733451843,
388
+ "learning_rate": 9.851478631379982e-05,
389
+ "loss": 0.9038,
390
+ "step": 52
391
+ },
392
+ {
393
+ "epoch": 0.45250800426894344,
394
+ "grad_norm": 0.30273666977882385,
395
+ "learning_rate": 9.837076097314319e-05,
396
+ "loss": 0.9348,
397
+ "step": 53
398
+ },
399
+ {
400
+ "epoch": 0.4610458911419424,
401
+ "grad_norm": 0.3075931966304779,
402
+ "learning_rate": 9.822018700574695e-05,
403
+ "loss": 0.9275,
404
+ "step": 54
405
+ },
406
+ {
407
+ "epoch": 0.4695837780149413,
408
+ "grad_norm": 0.2986295521259308,
409
+ "learning_rate": 9.806308479691595e-05,
410
+ "loss": 0.928,
411
+ "step": 55
412
+ },
413
+ {
414
+ "epoch": 0.4781216648879402,
415
+ "grad_norm": 0.3007640540599823,
416
+ "learning_rate": 9.789947561577445e-05,
417
+ "loss": 0.9702,
418
+ "step": 56
419
+ },
420
+ {
421
+ "epoch": 0.48665955176093917,
422
+ "grad_norm": 0.30233699083328247,
423
+ "learning_rate": 9.77293816123866e-05,
424
+ "loss": 0.9534,
425
+ "step": 57
426
+ },
427
+ {
428
+ "epoch": 0.4951974386339381,
429
+ "grad_norm": 0.30345094203948975,
430
+ "learning_rate": 9.755282581475769e-05,
431
+ "loss": 0.9474,
432
+ "step": 58
433
+ },
434
+ {
435
+ "epoch": 0.503735325506937,
436
+ "grad_norm": 0.3086802065372467,
437
+ "learning_rate": 9.736983212571646e-05,
438
+ "loss": 0.8856,
439
+ "step": 59
440
+ },
441
+ {
442
+ "epoch": 0.512273212379936,
443
+ "grad_norm": 0.28710058331489563,
444
+ "learning_rate": 9.718042531967918e-05,
445
+ "loss": 0.946,
446
+ "step": 60
447
+ },
448
+ {
449
+ "epoch": 0.5208110992529349,
450
+ "grad_norm": 0.3032543361186981,
451
+ "learning_rate": 9.698463103929542e-05,
452
+ "loss": 0.9204,
453
+ "step": 61
454
+ },
455
+ {
456
+ "epoch": 0.5293489861259338,
457
+ "grad_norm": 0.33324557542800903,
458
+ "learning_rate": 9.678247579197657e-05,
459
+ "loss": 0.9127,
460
+ "step": 62
461
+ },
462
+ {
463
+ "epoch": 0.5378868729989328,
464
+ "grad_norm": 0.31016331911087036,
465
+ "learning_rate": 9.657398694630712e-05,
466
+ "loss": 0.8976,
467
+ "step": 63
468
+ },
469
+ {
470
+ "epoch": 0.5464247598719317,
471
+ "grad_norm": 0.3162042498588562,
472
+ "learning_rate": 9.635919272833938e-05,
473
+ "loss": 0.8394,
474
+ "step": 64
475
+ },
476
+ {
477
+ "epoch": 0.5549626467449307,
478
+ "grad_norm": 0.3050558865070343,
479
+ "learning_rate": 9.613812221777212e-05,
480
+ "loss": 0.9153,
481
+ "step": 65
482
+ },
483
+ {
484
+ "epoch": 0.5635005336179295,
485
+ "grad_norm": 0.3125409483909607,
486
+ "learning_rate": 9.591080534401371e-05,
487
+ "loss": 0.9224,
488
+ "step": 66
489
+ },
490
+ {
491
+ "epoch": 0.5720384204909285,
492
+ "grad_norm": 0.3080565333366394,
493
+ "learning_rate": 9.567727288213005e-05,
494
+ "loss": 0.9281,
495
+ "step": 67
496
+ },
497
+ {
498
+ "epoch": 0.5805763073639274,
499
+ "grad_norm": 0.31865012645721436,
500
+ "learning_rate": 9.543755644867822e-05,
501
+ "loss": 0.8459,
502
+ "step": 68
503
+ },
504
+ {
505
+ "epoch": 0.5891141942369263,
506
+ "grad_norm": 0.3252212107181549,
507
+ "learning_rate": 9.519168849742604e-05,
508
+ "loss": 0.9076,
509
+ "step": 69
510
+ },
511
+ {
512
+ "epoch": 0.5976520811099253,
513
+ "grad_norm": 0.31885048747062683,
514
+ "learning_rate": 9.493970231495835e-05,
515
+ "loss": 0.9292,
516
+ "step": 70
517
+ },
518
+ {
519
+ "epoch": 0.6061899679829242,
520
+ "grad_norm": 0.3276972770690918,
521
+ "learning_rate": 9.468163201617062e-05,
522
+ "loss": 0.7814,
523
+ "step": 71
524
+ },
525
+ {
526
+ "epoch": 0.6147278548559232,
527
+ "grad_norm": 0.3108748495578766,
528
+ "learning_rate": 9.441751253965021e-05,
529
+ "loss": 0.8878,
530
+ "step": 72
531
+ },
532
+ {
533
+ "epoch": 0.6232657417289221,
534
+ "grad_norm": 0.3175681531429291,
535
+ "learning_rate": 9.414737964294636e-05,
536
+ "loss": 0.9064,
537
+ "step": 73
538
+ },
539
+ {
540
+ "epoch": 0.6318036286019211,
541
+ "grad_norm": 0.3110436499118805,
542
+ "learning_rate": 9.38712698977291e-05,
543
+ "loss": 0.9185,
544
+ "step": 74
545
+ },
546
+ {
547
+ "epoch": 0.6403415154749199,
548
+ "grad_norm": 0.31339648365974426,
549
+ "learning_rate": 9.358922068483812e-05,
550
+ "loss": 0.9435,
551
+ "step": 75
552
+ },
553
+ {
554
+ "epoch": 0.6403415154749199,
555
+ "eval_loss": 0.8910433650016785,
556
+ "eval_runtime": 126.3701,
557
+ "eval_samples_per_second": 7.739,
558
+ "eval_steps_per_second": 0.491,
559
+ "step": 75
560
+ },
561
+ {
562
+ "epoch": 0.6488794023479189,
563
+ "grad_norm": 0.32126352190971375,
564
+ "learning_rate": 9.330127018922194e-05,
565
+ "loss": 0.91,
566
+ "step": 76
567
+ },
568
+ {
569
+ "epoch": 0.6574172892209178,
570
+ "grad_norm": 0.31005603075027466,
571
+ "learning_rate": 9.300745739476829e-05,
572
+ "loss": 0.9656,
573
+ "step": 77
574
+ },
575
+ {
576
+ "epoch": 0.6659551760939167,
577
+ "grad_norm": 0.3075394630432129,
578
+ "learning_rate": 9.270782207902629e-05,
579
+ "loss": 0.9734,
580
+ "step": 78
581
+ },
582
+ {
583
+ "epoch": 0.6744930629669157,
584
+ "grad_norm": 0.3072083294391632,
585
+ "learning_rate": 9.24024048078213e-05,
586
+ "loss": 0.9102,
587
+ "step": 79
588
+ },
589
+ {
590
+ "epoch": 0.6830309498399146,
591
+ "grad_norm": 0.3321980834007263,
592
+ "learning_rate": 9.209124692976287e-05,
593
+ "loss": 0.9166,
594
+ "step": 80
595
+ },
596
+ {
597
+ "epoch": 0.6915688367129136,
598
+ "grad_norm": 0.3266827464103699,
599
+ "learning_rate": 9.177439057064683e-05,
600
+ "loss": 0.8672,
601
+ "step": 81
602
+ },
603
+ {
604
+ "epoch": 0.7001067235859125,
605
+ "grad_norm": 0.33046409487724304,
606
+ "learning_rate": 9.145187862775209e-05,
607
+ "loss": 0.8953,
608
+ "step": 82
609
+ },
610
+ {
611
+ "epoch": 0.7086446104589115,
612
+ "grad_norm": 0.33084535598754883,
613
+ "learning_rate": 9.112375476403312e-05,
614
+ "loss": 0.914,
615
+ "step": 83
616
+ },
617
+ {
618
+ "epoch": 0.7171824973319103,
619
+ "grad_norm": 0.3533576726913452,
620
+ "learning_rate": 9.079006340220862e-05,
621
+ "loss": 0.8936,
622
+ "step": 84
623
+ },
624
+ {
625
+ "epoch": 0.7257203842049093,
626
+ "grad_norm": 0.3123546242713928,
627
+ "learning_rate": 9.045084971874738e-05,
628
+ "loss": 0.9096,
629
+ "step": 85
630
+ },
631
+ {
632
+ "epoch": 0.7342582710779082,
633
+ "grad_norm": 0.3312874138355255,
634
+ "learning_rate": 9.01061596377522e-05,
635
+ "loss": 0.957,
636
+ "step": 86
637
+ },
638
+ {
639
+ "epoch": 0.7427961579509071,
640
+ "grad_norm": 0.3258456885814667,
641
+ "learning_rate": 8.97560398247424e-05,
642
+ "loss": 0.8853,
643
+ "step": 87
644
+ },
645
+ {
646
+ "epoch": 0.7513340448239061,
647
+ "grad_norm": 0.35378894209861755,
648
+ "learning_rate": 8.940053768033609e-05,
649
+ "loss": 0.907,
650
+ "step": 88
651
+ },
652
+ {
653
+ "epoch": 0.759871931696905,
654
+ "grad_norm": 0.30806443095207214,
655
+ "learning_rate": 8.903970133383297e-05,
656
+ "loss": 0.8719,
657
+ "step": 89
658
+ },
659
+ {
660
+ "epoch": 0.768409818569904,
661
+ "grad_norm": 0.33278632164001465,
662
+ "learning_rate": 8.86735796366982e-05,
663
+ "loss": 0.9782,
664
+ "step": 90
665
+ },
666
+ {
667
+ "epoch": 0.7769477054429029,
668
+ "grad_norm": 0.31797873973846436,
669
+ "learning_rate": 8.83022221559489e-05,
670
+ "loss": 0.9225,
671
+ "step": 91
672
+ },
673
+ {
674
+ "epoch": 0.7854855923159018,
675
+ "grad_norm": 0.3217017352581024,
676
+ "learning_rate": 8.792567916744346e-05,
677
+ "loss": 0.9045,
678
+ "step": 92
679
+ },
680
+ {
681
+ "epoch": 0.7940234791889007,
682
+ "grad_norm": 0.37561750411987305,
683
+ "learning_rate": 8.754400164907497e-05,
684
+ "loss": 0.9335,
685
+ "step": 93
686
+ },
687
+ {
688
+ "epoch": 0.8025613660618997,
689
+ "grad_norm": 0.3165503144264221,
690
+ "learning_rate": 8.715724127386972e-05,
691
+ "loss": 0.9086,
692
+ "step": 94
693
+ },
694
+ {
695
+ "epoch": 0.8110992529348986,
696
+ "grad_norm": 0.3458029329776764,
697
+ "learning_rate": 8.676545040299145e-05,
698
+ "loss": 0.9079,
699
+ "step": 95
700
+ },
701
+ {
702
+ "epoch": 0.8196371398078975,
703
+ "grad_norm": 0.336024671792984,
704
+ "learning_rate": 8.636868207865244e-05,
705
+ "loss": 0.9059,
706
+ "step": 96
707
+ },
708
+ {
709
+ "epoch": 0.8281750266808965,
710
+ "grad_norm": 0.3598126471042633,
711
+ "learning_rate": 8.596699001693255e-05,
712
+ "loss": 0.847,
713
+ "step": 97
714
+ },
715
+ {
716
+ "epoch": 0.8367129135538954,
717
+ "grad_norm": 0.33313530683517456,
718
+ "learning_rate": 8.556042860050687e-05,
719
+ "loss": 0.911,
720
+ "step": 98
721
+ },
722
+ {
723
+ "epoch": 0.8452508004268944,
724
+ "grad_norm": 0.3099718689918518,
725
+ "learning_rate": 8.51490528712831e-05,
726
+ "loss": 0.8864,
727
+ "step": 99
728
+ },
729
+ {
730
+ "epoch": 0.8537886872998933,
731
+ "grad_norm": 0.3349950611591339,
732
+ "learning_rate": 8.473291852294987e-05,
733
+ "loss": 0.8479,
734
+ "step": 100
735
+ },
736
+ {
737
+ "epoch": 0.8537886872998933,
738
+ "eval_loss": 0.8699346780776978,
739
+ "eval_runtime": 126.3565,
740
+ "eval_samples_per_second": 7.74,
741
+ "eval_steps_per_second": 0.491,
742
+ "step": 100
743
+ },
744
+ {
745
+ "epoch": 0.8623265741728922,
746
+ "grad_norm": 0.3409707546234131,
747
+ "learning_rate": 8.43120818934367e-05,
748
+ "loss": 0.92,
749
+ "step": 101
750
+ },
751
+ {
752
+ "epoch": 0.8708644610458911,
753
+ "grad_norm": 0.349369078874588,
754
+ "learning_rate": 8.388659995728663e-05,
755
+ "loss": 0.8551,
756
+ "step": 102
757
+ },
758
+ {
759
+ "epoch": 0.87940234791889,
760
+ "grad_norm": 0.3358254134654999,
761
+ "learning_rate": 8.345653031794292e-05,
762
+ "loss": 0.9024,
763
+ "step": 103
764
+ },
765
+ {
766
+ "epoch": 0.887940234791889,
767
+ "grad_norm": 0.30185699462890625,
768
+ "learning_rate": 8.302193119995039e-05,
769
+ "loss": 0.9081,
770
+ "step": 104
771
+ },
772
+ {
773
+ "epoch": 0.8964781216648879,
774
+ "grad_norm": 0.33944496512413025,
775
+ "learning_rate": 8.258286144107276e-05,
776
+ "loss": 0.9215,
777
+ "step": 105
778
+ },
779
+ {
780
+ "epoch": 0.9050160085378869,
781
+ "grad_norm": 0.31531962752342224,
782
+ "learning_rate": 8.213938048432697e-05,
783
+ "loss": 0.886,
784
+ "step": 106
785
+ },
786
+ {
787
+ "epoch": 0.9135538954108858,
788
+ "grad_norm": 0.33005067706108093,
789
+ "learning_rate": 8.169154836993551e-05,
790
+ "loss": 0.8436,
791
+ "step": 107
792
+ },
793
+ {
794
+ "epoch": 0.9220917822838848,
795
+ "grad_norm": 0.32761579751968384,
796
+ "learning_rate": 8.1239425727198e-05,
797
+ "loss": 0.8181,
798
+ "step": 108
799
+ },
800
+ {
801
+ "epoch": 0.9306296691568837,
802
+ "grad_norm": 0.3269185423851013,
803
+ "learning_rate": 8.07830737662829e-05,
804
+ "loss": 0.8687,
805
+ "step": 109
806
+ },
807
+ {
808
+ "epoch": 0.9391675560298826,
809
+ "grad_norm": 0.3282541036605835,
810
+ "learning_rate": 8.032255426994069e-05,
811
+ "loss": 0.8341,
812
+ "step": 110
813
+ },
814
+ {
815
+ "epoch": 0.9477054429028815,
816
+ "grad_norm": 0.331903338432312,
817
+ "learning_rate": 7.985792958513931e-05,
818
+ "loss": 0.9307,
819
+ "step": 111
820
+ },
821
+ {
822
+ "epoch": 0.9562433297758804,
823
+ "grad_norm": 0.37399402260780334,
824
+ "learning_rate": 7.938926261462366e-05,
825
+ "loss": 0.8765,
826
+ "step": 112
827
+ },
828
+ {
829
+ "epoch": 0.9647812166488794,
830
+ "grad_norm": 0.3470500409603119,
831
+ "learning_rate": 7.891661680839932e-05,
832
+ "loss": 0.8956,
833
+ "step": 113
834
+ },
835
+ {
836
+ "epoch": 0.9733191035218783,
837
+ "grad_norm": 0.33433717489242554,
838
+ "learning_rate": 7.844005615514259e-05,
839
+ "loss": 0.8607,
840
+ "step": 114
841
+ },
842
+ {
843
+ "epoch": 0.9818569903948773,
844
+ "grad_norm": 0.3233843743801117,
845
+ "learning_rate": 7.795964517353735e-05,
846
+ "loss": 0.8357,
847
+ "step": 115
848
+ },
849
+ {
850
+ "epoch": 0.9903948772678762,
851
+ "grad_norm": 0.32620981335639954,
852
+ "learning_rate": 7.74754489035403e-05,
853
+ "loss": 0.9093,
854
+ "step": 116
855
+ },
856
+ {
857
+ "epoch": 0.9989327641408752,
858
+ "grad_norm": 0.3210049569606781,
859
+ "learning_rate": 7.698753289757565e-05,
860
+ "loss": 0.8595,
861
+ "step": 117
862
+ },
863
+ {
864
+ "epoch": 1.0,
865
+ "grad_norm": 0.883870542049408,
866
+ "learning_rate": 7.649596321166024e-05,
867
+ "loss": 0.968,
868
+ "step": 118
869
+ },
870
+ {
871
+ "epoch": 1.0085378868729988,
872
+ "grad_norm": 0.36958256363868713,
873
+ "learning_rate": 7.600080639646077e-05,
874
+ "loss": 0.8389,
875
+ "step": 119
876
+ },
877
+ {
878
+ "epoch": 1.017075773745998,
879
+ "grad_norm": 0.32345816493034363,
880
+ "learning_rate": 7.550212948828377e-05,
881
+ "loss": 0.8826,
882
+ "step": 120
883
+ },
884
+ {
885
+ "epoch": 1.0256136606189967,
886
+ "grad_norm": 0.33251067996025085,
887
+ "learning_rate": 7.500000000000001e-05,
888
+ "loss": 0.8961,
889
+ "step": 121
890
+ },
891
+ {
892
+ "epoch": 1.0341515474919958,
893
+ "grad_norm": 0.3673892915248871,
894
+ "learning_rate": 7.449448591190435e-05,
895
+ "loss": 0.7779,
896
+ "step": 122
897
+ },
898
+ {
899
+ "epoch": 1.0426894343649946,
900
+ "grad_norm": 0.3406680226325989,
901
+ "learning_rate": 7.398565566251232e-05,
902
+ "loss": 0.882,
903
+ "step": 123
904
+ },
905
+ {
906
+ "epoch": 1.0512273212379937,
907
+ "grad_norm": 0.35136982798576355,
908
+ "learning_rate": 7.347357813929454e-05,
909
+ "loss": 0.8583,
910
+ "step": 124
911
+ },
912
+ {
913
+ "epoch": 1.0597652081109925,
914
+ "grad_norm": 0.38215872645378113,
915
+ "learning_rate": 7.295832266935059e-05,
916
+ "loss": 0.8085,
917
+ "step": 125
918
+ },
919
+ {
920
+ "epoch": 1.0597652081109925,
921
+ "eval_loss": 0.855903685092926,
922
+ "eval_runtime": 126.4753,
923
+ "eval_samples_per_second": 7.733,
924
+ "eval_steps_per_second": 0.49,
925
+ "step": 125
926
+ },
927
+ {
928
+ "epoch": 1.0683030949839916,
929
+ "grad_norm": 0.3545176386833191,
930
+ "learning_rate": 7.243995901002312e-05,
931
+ "loss": 0.8287,
932
+ "step": 126
933
+ },
934
+ {
935
+ "epoch": 1.0768409818569904,
936
+ "grad_norm": 0.36242854595184326,
937
+ "learning_rate": 7.191855733945387e-05,
938
+ "loss": 0.8779,
939
+ "step": 127
940
+ },
941
+ {
942
+ "epoch": 1.0853788687299892,
943
+ "grad_norm": 0.3716667890548706,
944
+ "learning_rate": 7.139418824708272e-05,
945
+ "loss": 0.8207,
946
+ "step": 128
947
+ },
948
+ {
949
+ "epoch": 1.0939167556029883,
950
+ "grad_norm": 0.3446180820465088,
951
+ "learning_rate": 7.08669227240909e-05,
952
+ "loss": 0.8071,
953
+ "step": 129
954
+ },
955
+ {
956
+ "epoch": 1.1024546424759871,
957
+ "grad_norm": 0.348202109336853,
958
+ "learning_rate": 7.033683215379002e-05,
959
+ "loss": 0.816,
960
+ "step": 130
961
+ },
962
+ {
963
+ "epoch": 1.1109925293489862,
964
+ "grad_norm": 0.3857647776603699,
965
+ "learning_rate": 6.980398830195785e-05,
966
+ "loss": 0.8307,
967
+ "step": 131
968
+ },
969
+ {
970
+ "epoch": 1.119530416221985,
971
+ "grad_norm": 0.3619552254676819,
972
+ "learning_rate": 6.926846330712242e-05,
973
+ "loss": 0.8799,
974
+ "step": 132
975
+ },
976
+ {
977
+ "epoch": 1.128068303094984,
978
+ "grad_norm": 0.36076459288597107,
979
+ "learning_rate": 6.873032967079561e-05,
980
+ "loss": 0.8427,
981
+ "step": 133
982
+ },
983
+ {
984
+ "epoch": 1.136606189967983,
985
+ "grad_norm": 0.3395594358444214,
986
+ "learning_rate": 6.818966024765758e-05,
987
+ "loss": 0.857,
988
+ "step": 134
989
+ },
990
+ {
991
+ "epoch": 1.1451440768409817,
992
+ "grad_norm": 0.3523291051387787,
993
+ "learning_rate": 6.764652823569344e-05,
994
+ "loss": 0.8238,
995
+ "step": 135
996
+ },
997
+ {
998
+ "epoch": 1.1536819637139808,
999
+ "grad_norm": 0.35138288140296936,
1000
+ "learning_rate": 6.710100716628344e-05,
1001
+ "loss": 0.8208,
1002
+ "step": 136
1003
+ },
1004
+ {
1005
+ "epoch": 1.1622198505869796,
1006
+ "grad_norm": 0.3666648864746094,
1007
+ "learning_rate": 6.65531708942479e-05,
1008
+ "loss": 0.7936,
1009
+ "step": 137
1010
+ },
1011
+ {
1012
+ "epoch": 1.1707577374599787,
1013
+ "grad_norm": 0.38198062777519226,
1014
+ "learning_rate": 6.600309358784857e-05,
1015
+ "loss": 0.8355,
1016
+ "step": 138
1017
+ },
1018
+ {
1019
+ "epoch": 1.1792956243329775,
1020
+ "grad_norm": 0.361715167760849,
1021
+ "learning_rate": 6.545084971874738e-05,
1022
+ "loss": 0.7518,
1023
+ "step": 139
1024
+ },
1025
+ {
1026
+ "epoch": 1.1878335112059766,
1027
+ "grad_norm": 0.3789786100387573,
1028
+ "learning_rate": 6.48965140519241e-05,
1029
+ "loss": 0.8091,
1030
+ "step": 140
1031
+ },
1032
+ {
1033
+ "epoch": 1.1963713980789754,
1034
+ "grad_norm": 0.3489861488342285,
1035
+ "learning_rate": 6.434016163555452e-05,
1036
+ "loss": 0.8138,
1037
+ "step": 141
1038
+ },
1039
+ {
1040
+ "epoch": 1.2049092849519745,
1041
+ "grad_norm": 0.3795143961906433,
1042
+ "learning_rate": 6.378186779084995e-05,
1043
+ "loss": 0.8357,
1044
+ "step": 142
1045
+ },
1046
+ {
1047
+ "epoch": 1.2134471718249733,
1048
+ "grad_norm": 0.3954067826271057,
1049
+ "learning_rate": 6.322170810186012e-05,
1050
+ "loss": 0.8278,
1051
+ "step": 143
1052
+ },
1053
+ {
1054
+ "epoch": 1.2219850586979724,
1055
+ "grad_norm": 0.37652266025543213,
1056
+ "learning_rate": 6.26597584052401e-05,
1057
+ "loss": 0.8395,
1058
+ "step": 144
1059
+ },
1060
+ {
1061
+ "epoch": 1.2305229455709712,
1062
+ "grad_norm": 0.3819694519042969,
1063
+ "learning_rate": 6.209609477998338e-05,
1064
+ "loss": 0.811,
1065
+ "step": 145
1066
+ },
1067
+ {
1068
+ "epoch": 1.23906083244397,
1069
+ "grad_norm": 0.3565637767314911,
1070
+ "learning_rate": 6.153079353712201e-05,
1071
+ "loss": 0.8305,
1072
+ "step": 146
1073
+ },
1074
+ {
1075
+ "epoch": 1.247598719316969,
1076
+ "grad_norm": 0.37492281198501587,
1077
+ "learning_rate": 6.096393120939516e-05,
1078
+ "loss": 0.8235,
1079
+ "step": 147
1080
+ },
1081
+ {
1082
+ "epoch": 1.256136606189968,
1083
+ "grad_norm": 0.3353641927242279,
1084
+ "learning_rate": 6.0395584540887963e-05,
1085
+ "loss": 0.8181,
1086
+ "step": 148
1087
+ },
1088
+ {
1089
+ "epoch": 1.264674493062967,
1090
+ "grad_norm": 0.37503373622894287,
1091
+ "learning_rate": 5.982583047664151e-05,
1092
+ "loss": 0.8309,
1093
+ "step": 149
1094
+ },
1095
+ {
1096
+ "epoch": 1.2732123799359658,
1097
+ "grad_norm": 0.34756171703338623,
1098
+ "learning_rate": 5.925474615223573e-05,
1099
+ "loss": 0.8345,
1100
+ "step": 150
1101
+ },
1102
+ {
1103
+ "epoch": 1.2732123799359658,
1104
+ "eval_loss": 0.8430656790733337,
1105
+ "eval_runtime": 126.4713,
1106
+ "eval_samples_per_second": 7.733,
1107
+ "eval_steps_per_second": 0.49,
1108
+ "step": 150
1109
+ }
1110
+ ],
1111
+ "logging_steps": 1,
1112
+ "max_steps": 300,
1113
+ "num_input_tokens_seen": 0,
1114
+ "num_train_epochs": 3,
1115
+ "save_steps": 50,
1116
+ "stateful_callbacks": {
1117
+ "TrainerControl": {
1118
+ "args": {
1119
+ "should_epoch_stop": false,
1120
+ "should_evaluate": false,
1121
+ "should_log": false,
1122
+ "should_save": true,
1123
+ "should_training_stop": false
1124
+ },
1125
+ "attributes": {}
1126
+ }
1127
+ },
1128
+ "total_flos": 1.4140777917830922e+18,
1129
+ "train_batch_size": 3,
1130
+ "trial_name": null,
1131
+ "trial_params": null
1132
+ }
TextUI-Trans-7B/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:317d8fbe31bcf06a0ec571c13c58278b73ace8dc6147ca47db99a7934571c640
3
+ size 6161
TextUI-Trans-7B/vocab.json ADDED
The diff for this file is too large to render. See raw diff