sang-w00 commited on
Commit
e47d5f9
·
verified ·
1 Parent(s): 3bf357b

Delete sft-clevr-1106-20epoch/checkpoint-2500

Browse files
sft-clevr-1106-20epoch/checkpoint-2500/added_tokens.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "</tool_call>": 151658,
3
- "<tool_call>": 151657,
4
- "<|box_end|>": 151649,
5
- "<|box_start|>": 151648,
6
- "<|endoftext|>": 151643,
7
- "<|file_sep|>": 151664,
8
- "<|fim_middle|>": 151660,
9
- "<|fim_pad|>": 151662,
10
- "<|fim_prefix|>": 151659,
11
- "<|fim_suffix|>": 151661,
12
- "<|im_end|>": 151645,
13
- "<|im_start|>": 151644,
14
- "<|image_pad|>": 151655,
15
- "<|object_ref_end|>": 151647,
16
- "<|object_ref_start|>": 151646,
17
- "<|quad_end|>": 151651,
18
- "<|quad_start|>": 151650,
19
- "<|repo_name|>": 151663,
20
- "<|video_pad|>": 151656,
21
- "<|vision_end|>": 151653,
22
- "<|vision_pad|>": 151654,
23
- "<|vision_start|>": 151652
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/chat_template.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
3
- }
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/config.json DELETED
@@ -1,50 +0,0 @@
1
- {
2
- "_name_or_path": "Qwen/Qwen2.5-VL-7B-Instruct",
3
- "architectures": [
4
- "Qwen2_5_VLForConditionalGeneration"
5
- ],
6
- "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
- "eos_token_id": 151645,
9
- "hidden_act": "silu",
10
- "hidden_size": 3584,
11
- "image_token_id": 151655,
12
- "initializer_range": 0.02,
13
- "intermediate_size": 18944,
14
- "max_position_embeddings": 128000,
15
- "max_window_layers": 28,
16
- "model_type": "qwen2_5_vl",
17
- "num_attention_heads": 28,
18
- "num_hidden_layers": 28,
19
- "num_key_value_heads": 4,
20
- "rms_norm_eps": 1e-06,
21
- "rope_scaling": {
22
- "mrope_section": [
23
- 16,
24
- 24,
25
- 24
26
- ],
27
- "rope_type": "default",
28
- "type": "default"
29
- },
30
- "rope_theta": 1000000.0,
31
- "sliding_window": 32768,
32
- "tie_word_embeddings": false,
33
- "torch_dtype": "bfloat16",
34
- "transformers_version": "4.49.0",
35
- "use_cache": false,
36
- "use_sliding_window": false,
37
- "video_token_id": 151656,
38
- "vision_config": {
39
- "hidden_size": 1280,
40
- "in_chans": 3,
41
- "model_type": "qwen2_5_vl",
42
- "spatial_patch_size": 14,
43
- "tokens_per_second": 2,
44
- "torch_dtype": "bfloat16"
45
- },
46
- "vision_end_token_id": 151653,
47
- "vision_start_token_id": 151652,
48
- "vision_token_id": 151654,
49
- "vocab_size": 152064
50
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/generation_config.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "attn_implementation": "flash_attention_2",
3
- "bos_token_id": 151643,
4
- "do_sample": true,
5
- "eos_token_id": [
6
- 151645,
7
- 151643
8
- ],
9
- "pad_token_id": 151643,
10
- "repetition_penalty": 1.05,
11
- "temperature": 1e-06,
12
- "transformers_version": "4.49.0",
13
- "use_cache": false
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
sft-clevr-1106-20epoch/checkpoint-2500/model-00001-of-00004.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f5a41d63e65d883f59ebb4573c9f00a2186c601607308378ad14493bcf18a29
3
- size 4968243304
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/model-00002-of-00004.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e04d5d29b50ed607f477415ba0078d77d16558454127087b34c2708c7becd93
3
- size 4991495816
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/model-00003-of-00004.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:69bee98a05456798942df0fc5e3e7558cf9d0ce29b6f3032ef0691affcecef18
3
- size 4932751040
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/model-00004-of-00004.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:135d6c5ef42de9fe78889c22fd56e5c61719e41d757881e8b1c0de06554ac165
3
- size 1691924384
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/model.safetensors.index.json DELETED
@@ -1,736 +0,0 @@
1
- {
2
- "metadata": {
3
- "total_size": 16584333312
4
- },
5
- "weight_map": {
6
- "lm_head.weight": "model-00004-of-00004.safetensors",
7
- "model.embed_tokens.weight": "model-00001-of-00004.safetensors",
8
- "model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
9
- "model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
10
- "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
11
- "model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
12
- "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
13
- "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
14
- "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
15
- "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
16
- "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
17
- "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
18
- "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
19
- "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
20
- "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
21
- "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
22
- "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
23
- "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
24
- "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
25
- "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
26
- "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
27
- "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
28
- "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
29
- "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
30
- "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
31
- "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
32
- "model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
33
- "model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
34
- "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
35
- "model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
36
- "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
37
- "model.layers.10.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
38
- "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
39
- "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
40
- "model.layers.10.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
41
- "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
42
- "model.layers.10.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
43
- "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
44
- "model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
45
- "model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
46
- "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
47
- "model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
48
- "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
49
- "model.layers.11.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
50
- "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
51
- "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
52
- "model.layers.11.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
53
- "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
54
- "model.layers.11.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
55
- "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
56
- "model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
57
- "model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
58
- "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
59
- "model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
60
- "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
61
- "model.layers.12.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
62
- "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
63
- "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
64
- "model.layers.12.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
65
- "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
66
- "model.layers.12.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
67
- "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
68
- "model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
69
- "model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
70
- "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
71
- "model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
72
- "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
73
- "model.layers.13.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
74
- "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
75
- "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
76
- "model.layers.13.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
77
- "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
78
- "model.layers.13.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
79
- "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
80
- "model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
81
- "model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
82
- "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
83
- "model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
84
- "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
85
- "model.layers.14.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
86
- "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
87
- "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
88
- "model.layers.14.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
89
- "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
90
- "model.layers.14.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
91
- "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
92
- "model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
93
- "model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
94
- "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
95
- "model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
96
- "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
97
- "model.layers.15.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
98
- "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
99
- "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
100
- "model.layers.15.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
101
- "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
102
- "model.layers.15.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
103
- "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
104
- "model.layers.16.input_layernorm.weight": "model-00003-of-00004.safetensors",
105
- "model.layers.16.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
106
- "model.layers.16.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
107
- "model.layers.16.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
108
- "model.layers.16.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
109
- "model.layers.16.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
110
- "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
111
- "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
112
- "model.layers.16.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
113
- "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
114
- "model.layers.16.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
115
- "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
116
- "model.layers.17.input_layernorm.weight": "model-00003-of-00004.safetensors",
117
- "model.layers.17.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
118
- "model.layers.17.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
119
- "model.layers.17.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
120
- "model.layers.17.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
121
- "model.layers.17.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
122
- "model.layers.17.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
123
- "model.layers.17.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
124
- "model.layers.17.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
125
- "model.layers.17.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
126
- "model.layers.17.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
127
- "model.layers.17.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
128
- "model.layers.18.input_layernorm.weight": "model-00003-of-00004.safetensors",
129
- "model.layers.18.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
130
- "model.layers.18.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
131
- "model.layers.18.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
132
- "model.layers.18.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
133
- "model.layers.18.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
134
- "model.layers.18.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
135
- "model.layers.18.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
136
- "model.layers.18.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
137
- "model.layers.18.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
138
- "model.layers.18.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
139
- "model.layers.18.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
140
- "model.layers.19.input_layernorm.weight": "model-00003-of-00004.safetensors",
141
- "model.layers.19.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
142
- "model.layers.19.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
143
- "model.layers.19.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
144
- "model.layers.19.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
145
- "model.layers.19.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
146
- "model.layers.19.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
147
- "model.layers.19.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
148
- "model.layers.19.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
149
- "model.layers.19.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
150
- "model.layers.19.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
151
- "model.layers.19.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
152
- "model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
153
- "model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
154
- "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
155
- "model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
156
- "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
157
- "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
158
- "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
159
- "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
160
- "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
161
- "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
162
- "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
163
- "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
164
- "model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors",
165
- "model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
166
- "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
167
- "model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
168
- "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
169
- "model.layers.20.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
170
- "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
171
- "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
172
- "model.layers.20.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
173
- "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
174
- "model.layers.20.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
175
- "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
176
- "model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors",
177
- "model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
178
- "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
179
- "model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
180
- "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
181
- "model.layers.21.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
182
- "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
183
- "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
184
- "model.layers.21.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
185
- "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
186
- "model.layers.21.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
187
- "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
188
- "model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
189
- "model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
190
- "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
191
- "model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
192
- "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
193
- "model.layers.22.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
194
- "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
195
- "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
196
- "model.layers.22.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
197
- "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
198
- "model.layers.22.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
199
- "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
200
- "model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors",
201
- "model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
202
- "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
203
- "model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
204
- "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
205
- "model.layers.23.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
206
- "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
207
- "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
208
- "model.layers.23.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
209
- "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
210
- "model.layers.23.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
211
- "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
212
- "model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors",
213
- "model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
214
- "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
215
- "model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
216
- "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
217
- "model.layers.24.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
218
- "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
219
- "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
220
- "model.layers.24.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
221
- "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
222
- "model.layers.24.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
223
- "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
224
- "model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
225
- "model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
226
- "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
227
- "model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
228
- "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
229
- "model.layers.25.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
230
- "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
231
- "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
232
- "model.layers.25.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
233
- "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
234
- "model.layers.25.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
235
- "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
236
- "model.layers.26.input_layernorm.weight": "model-00004-of-00004.safetensors",
237
- "model.layers.26.mlp.down_proj.weight": "model-00004-of-00004.safetensors",
238
- "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
239
- "model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
240
- "model.layers.26.post_attention_layernorm.weight": "model-00004-of-00004.safetensors",
241
- "model.layers.26.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
242
- "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
243
- "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
244
- "model.layers.26.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
245
- "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
246
- "model.layers.26.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
247
- "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
248
- "model.layers.27.input_layernorm.weight": "model-00004-of-00004.safetensors",
249
- "model.layers.27.mlp.down_proj.weight": "model-00004-of-00004.safetensors",
250
- "model.layers.27.mlp.gate_proj.weight": "model-00004-of-00004.safetensors",
251
- "model.layers.27.mlp.up_proj.weight": "model-00004-of-00004.safetensors",
252
- "model.layers.27.post_attention_layernorm.weight": "model-00004-of-00004.safetensors",
253
- "model.layers.27.self_attn.k_proj.bias": "model-00004-of-00004.safetensors",
254
- "model.layers.27.self_attn.k_proj.weight": "model-00004-of-00004.safetensors",
255
- "model.layers.27.self_attn.o_proj.weight": "model-00004-of-00004.safetensors",
256
- "model.layers.27.self_attn.q_proj.bias": "model-00004-of-00004.safetensors",
257
- "model.layers.27.self_attn.q_proj.weight": "model-00004-of-00004.safetensors",
258
- "model.layers.27.self_attn.v_proj.bias": "model-00004-of-00004.safetensors",
259
- "model.layers.27.self_attn.v_proj.weight": "model-00004-of-00004.safetensors",
260
- "model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
261
- "model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
262
- "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
263
- "model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
264
- "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
265
- "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
266
- "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
267
- "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
268
- "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
269
- "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
270
- "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
271
- "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
272
- "model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
273
- "model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
274
- "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
275
- "model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
276
- "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
277
- "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
278
- "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
279
- "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
280
- "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
281
- "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
282
- "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
283
- "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
284
- "model.layers.5.input_layernorm.weight": "model-00002-of-00004.safetensors",
285
- "model.layers.5.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
286
- "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
287
- "model.layers.5.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
288
- "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
289
- "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
290
- "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
291
- "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
292
- "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
293
- "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
294
- "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
295
- "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
296
- "model.layers.6.input_layernorm.weight": "model-00002-of-00004.safetensors",
297
- "model.layers.6.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
298
- "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
299
- "model.layers.6.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
300
- "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
301
- "model.layers.6.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
302
- "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
303
- "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
304
- "model.layers.6.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
305
- "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
306
- "model.layers.6.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
307
- "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
308
- "model.layers.7.input_layernorm.weight": "model-00002-of-00004.safetensors",
309
- "model.layers.7.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
310
- "model.layers.7.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
311
- "model.layers.7.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
312
- "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
313
- "model.layers.7.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
314
- "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
315
- "model.layers.7.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
316
- "model.layers.7.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
317
- "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
318
- "model.layers.7.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
319
- "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
320
- "model.layers.8.input_layernorm.weight": "model-00002-of-00004.safetensors",
321
- "model.layers.8.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
322
- "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
323
- "model.layers.8.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
324
- "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
325
- "model.layers.8.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
326
- "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
327
- "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
328
- "model.layers.8.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
329
- "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
330
- "model.layers.8.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
331
- "model.layers.8.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
332
- "model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
333
- "model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
334
- "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
335
- "model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
336
- "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
337
- "model.layers.9.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
338
- "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
339
- "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
340
- "model.layers.9.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
341
- "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
342
- "model.layers.9.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
343
- "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
344
- "model.norm.weight": "model-00004-of-00004.safetensors",
345
- "visual.blocks.0.attn.proj.bias": "model-00001-of-00004.safetensors",
346
- "visual.blocks.0.attn.proj.weight": "model-00001-of-00004.safetensors",
347
- "visual.blocks.0.attn.qkv.bias": "model-00001-of-00004.safetensors",
348
- "visual.blocks.0.attn.qkv.weight": "model-00001-of-00004.safetensors",
349
- "visual.blocks.0.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
350
- "visual.blocks.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
351
- "visual.blocks.0.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
352
- "visual.blocks.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
353
- "visual.blocks.0.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
354
- "visual.blocks.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
355
- "visual.blocks.0.norm1.weight": "model-00001-of-00004.safetensors",
356
- "visual.blocks.0.norm2.weight": "model-00001-of-00004.safetensors",
357
- "visual.blocks.1.attn.proj.bias": "model-00001-of-00004.safetensors",
358
- "visual.blocks.1.attn.proj.weight": "model-00001-of-00004.safetensors",
359
- "visual.blocks.1.attn.qkv.bias": "model-00001-of-00004.safetensors",
360
- "visual.blocks.1.attn.qkv.weight": "model-00001-of-00004.safetensors",
361
- "visual.blocks.1.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
362
- "visual.blocks.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
363
- "visual.blocks.1.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
364
- "visual.blocks.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
365
- "visual.blocks.1.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
366
- "visual.blocks.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
367
- "visual.blocks.1.norm1.weight": "model-00001-of-00004.safetensors",
368
- "visual.blocks.1.norm2.weight": "model-00001-of-00004.safetensors",
369
- "visual.blocks.10.attn.proj.bias": "model-00001-of-00004.safetensors",
370
- "visual.blocks.10.attn.proj.weight": "model-00001-of-00004.safetensors",
371
- "visual.blocks.10.attn.qkv.bias": "model-00001-of-00004.safetensors",
372
- "visual.blocks.10.attn.qkv.weight": "model-00001-of-00004.safetensors",
373
- "visual.blocks.10.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
374
- "visual.blocks.10.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
375
- "visual.blocks.10.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
376
- "visual.blocks.10.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
377
- "visual.blocks.10.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
378
- "visual.blocks.10.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
379
- "visual.blocks.10.norm1.weight": "model-00001-of-00004.safetensors",
380
- "visual.blocks.10.norm2.weight": "model-00001-of-00004.safetensors",
381
- "visual.blocks.11.attn.proj.bias": "model-00001-of-00004.safetensors",
382
- "visual.blocks.11.attn.proj.weight": "model-00001-of-00004.safetensors",
383
- "visual.blocks.11.attn.qkv.bias": "model-00001-of-00004.safetensors",
384
- "visual.blocks.11.attn.qkv.weight": "model-00001-of-00004.safetensors",
385
- "visual.blocks.11.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
386
- "visual.blocks.11.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
387
- "visual.blocks.11.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
388
- "visual.blocks.11.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
389
- "visual.blocks.11.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
390
- "visual.blocks.11.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
391
- "visual.blocks.11.norm1.weight": "model-00001-of-00004.safetensors",
392
- "visual.blocks.11.norm2.weight": "model-00001-of-00004.safetensors",
393
- "visual.blocks.12.attn.proj.bias": "model-00001-of-00004.safetensors",
394
- "visual.blocks.12.attn.proj.weight": "model-00001-of-00004.safetensors",
395
- "visual.blocks.12.attn.qkv.bias": "model-00001-of-00004.safetensors",
396
- "visual.blocks.12.attn.qkv.weight": "model-00001-of-00004.safetensors",
397
- "visual.blocks.12.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
398
- "visual.blocks.12.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
399
- "visual.blocks.12.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
400
- "visual.blocks.12.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
401
- "visual.blocks.12.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
402
- "visual.blocks.12.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
403
- "visual.blocks.12.norm1.weight": "model-00001-of-00004.safetensors",
404
- "visual.blocks.12.norm2.weight": "model-00001-of-00004.safetensors",
405
- "visual.blocks.13.attn.proj.bias": "model-00001-of-00004.safetensors",
406
- "visual.blocks.13.attn.proj.weight": "model-00001-of-00004.safetensors",
407
- "visual.blocks.13.attn.qkv.bias": "model-00001-of-00004.safetensors",
408
- "visual.blocks.13.attn.qkv.weight": "model-00001-of-00004.safetensors",
409
- "visual.blocks.13.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
410
- "visual.blocks.13.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
411
- "visual.blocks.13.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
412
- "visual.blocks.13.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
413
- "visual.blocks.13.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
414
- "visual.blocks.13.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
415
- "visual.blocks.13.norm1.weight": "model-00001-of-00004.safetensors",
416
- "visual.blocks.13.norm2.weight": "model-00001-of-00004.safetensors",
417
- "visual.blocks.14.attn.proj.bias": "model-00001-of-00004.safetensors",
418
- "visual.blocks.14.attn.proj.weight": "model-00001-of-00004.safetensors",
419
- "visual.blocks.14.attn.qkv.bias": "model-00001-of-00004.safetensors",
420
- "visual.blocks.14.attn.qkv.weight": "model-00001-of-00004.safetensors",
421
- "visual.blocks.14.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
422
- "visual.blocks.14.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
423
- "visual.blocks.14.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
424
- "visual.blocks.14.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
425
- "visual.blocks.14.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
426
- "visual.blocks.14.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
427
- "visual.blocks.14.norm1.weight": "model-00001-of-00004.safetensors",
428
- "visual.blocks.14.norm2.weight": "model-00001-of-00004.safetensors",
429
- "visual.blocks.15.attn.proj.bias": "model-00001-of-00004.safetensors",
430
- "visual.blocks.15.attn.proj.weight": "model-00001-of-00004.safetensors",
431
- "visual.blocks.15.attn.qkv.bias": "model-00001-of-00004.safetensors",
432
- "visual.blocks.15.attn.qkv.weight": "model-00001-of-00004.safetensors",
433
- "visual.blocks.15.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
434
- "visual.blocks.15.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
435
- "visual.blocks.15.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
436
- "visual.blocks.15.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
437
- "visual.blocks.15.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
438
- "visual.blocks.15.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
439
- "visual.blocks.15.norm1.weight": "model-00001-of-00004.safetensors",
440
- "visual.blocks.15.norm2.weight": "model-00001-of-00004.safetensors",
441
- "visual.blocks.16.attn.proj.bias": "model-00001-of-00004.safetensors",
442
- "visual.blocks.16.attn.proj.weight": "model-00001-of-00004.safetensors",
443
- "visual.blocks.16.attn.qkv.bias": "model-00001-of-00004.safetensors",
444
- "visual.blocks.16.attn.qkv.weight": "model-00001-of-00004.safetensors",
445
- "visual.blocks.16.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
446
- "visual.blocks.16.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
447
- "visual.blocks.16.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
448
- "visual.blocks.16.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
449
- "visual.blocks.16.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
450
- "visual.blocks.16.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
451
- "visual.blocks.16.norm1.weight": "model-00001-of-00004.safetensors",
452
- "visual.blocks.16.norm2.weight": "model-00001-of-00004.safetensors",
453
- "visual.blocks.17.attn.proj.bias": "model-00001-of-00004.safetensors",
454
- "visual.blocks.17.attn.proj.weight": "model-00001-of-00004.safetensors",
455
- "visual.blocks.17.attn.qkv.bias": "model-00001-of-00004.safetensors",
456
- "visual.blocks.17.attn.qkv.weight": "model-00001-of-00004.safetensors",
457
- "visual.blocks.17.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
458
- "visual.blocks.17.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
459
- "visual.blocks.17.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
460
- "visual.blocks.17.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
461
- "visual.blocks.17.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
462
- "visual.blocks.17.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
463
- "visual.blocks.17.norm1.weight": "model-00001-of-00004.safetensors",
464
- "visual.blocks.17.norm2.weight": "model-00001-of-00004.safetensors",
465
- "visual.blocks.18.attn.proj.bias": "model-00001-of-00004.safetensors",
466
- "visual.blocks.18.attn.proj.weight": "model-00001-of-00004.safetensors",
467
- "visual.blocks.18.attn.qkv.bias": "model-00001-of-00004.safetensors",
468
- "visual.blocks.18.attn.qkv.weight": "model-00001-of-00004.safetensors",
469
- "visual.blocks.18.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
470
- "visual.blocks.18.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
471
- "visual.blocks.18.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
472
- "visual.blocks.18.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
473
- "visual.blocks.18.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
474
- "visual.blocks.18.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
475
- "visual.blocks.18.norm1.weight": "model-00001-of-00004.safetensors",
476
- "visual.blocks.18.norm2.weight": "model-00001-of-00004.safetensors",
477
- "visual.blocks.19.attn.proj.bias": "model-00001-of-00004.safetensors",
478
- "visual.blocks.19.attn.proj.weight": "model-00001-of-00004.safetensors",
479
- "visual.blocks.19.attn.qkv.bias": "model-00001-of-00004.safetensors",
480
- "visual.blocks.19.attn.qkv.weight": "model-00001-of-00004.safetensors",
481
- "visual.blocks.19.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
482
- "visual.blocks.19.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
483
- "visual.blocks.19.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
484
- "visual.blocks.19.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
485
- "visual.blocks.19.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
486
- "visual.blocks.19.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
487
- "visual.blocks.19.norm1.weight": "model-00001-of-00004.safetensors",
488
- "visual.blocks.19.norm2.weight": "model-00001-of-00004.safetensors",
489
- "visual.blocks.2.attn.proj.bias": "model-00001-of-00004.safetensors",
490
- "visual.blocks.2.attn.proj.weight": "model-00001-of-00004.safetensors",
491
- "visual.blocks.2.attn.qkv.bias": "model-00001-of-00004.safetensors",
492
- "visual.blocks.2.attn.qkv.weight": "model-00001-of-00004.safetensors",
493
- "visual.blocks.2.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
494
- "visual.blocks.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
495
- "visual.blocks.2.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
496
- "visual.blocks.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
497
- "visual.blocks.2.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
498
- "visual.blocks.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
499
- "visual.blocks.2.norm1.weight": "model-00001-of-00004.safetensors",
500
- "visual.blocks.2.norm2.weight": "model-00001-of-00004.safetensors",
501
- "visual.blocks.20.attn.proj.bias": "model-00001-of-00004.safetensors",
502
- "visual.blocks.20.attn.proj.weight": "model-00001-of-00004.safetensors",
503
- "visual.blocks.20.attn.qkv.bias": "model-00001-of-00004.safetensors",
504
- "visual.blocks.20.attn.qkv.weight": "model-00001-of-00004.safetensors",
505
- "visual.blocks.20.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
506
- "visual.blocks.20.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
507
- "visual.blocks.20.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
508
- "visual.blocks.20.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
509
- "visual.blocks.20.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
510
- "visual.blocks.20.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
511
- "visual.blocks.20.norm1.weight": "model-00001-of-00004.safetensors",
512
- "visual.blocks.20.norm2.weight": "model-00001-of-00004.safetensors",
513
- "visual.blocks.21.attn.proj.bias": "model-00001-of-00004.safetensors",
514
- "visual.blocks.21.attn.proj.weight": "model-00001-of-00004.safetensors",
515
- "visual.blocks.21.attn.qkv.bias": "model-00001-of-00004.safetensors",
516
- "visual.blocks.21.attn.qkv.weight": "model-00001-of-00004.safetensors",
517
- "visual.blocks.21.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
518
- "visual.blocks.21.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
519
- "visual.blocks.21.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
520
- "visual.blocks.21.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
521
- "visual.blocks.21.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
522
- "visual.blocks.21.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
523
- "visual.blocks.21.norm1.weight": "model-00001-of-00004.safetensors",
524
- "visual.blocks.21.norm2.weight": "model-00001-of-00004.safetensors",
525
- "visual.blocks.22.attn.proj.bias": "model-00001-of-00004.safetensors",
526
- "visual.blocks.22.attn.proj.weight": "model-00001-of-00004.safetensors",
527
- "visual.blocks.22.attn.qkv.bias": "model-00001-of-00004.safetensors",
528
- "visual.blocks.22.attn.qkv.weight": "model-00001-of-00004.safetensors",
529
- "visual.blocks.22.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
530
- "visual.blocks.22.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
531
- "visual.blocks.22.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
532
- "visual.blocks.22.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
533
- "visual.blocks.22.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
534
- "visual.blocks.22.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
535
- "visual.blocks.22.norm1.weight": "model-00001-of-00004.safetensors",
536
- "visual.blocks.22.norm2.weight": "model-00001-of-00004.safetensors",
537
- "visual.blocks.23.attn.proj.bias": "model-00001-of-00004.safetensors",
538
- "visual.blocks.23.attn.proj.weight": "model-00001-of-00004.safetensors",
539
- "visual.blocks.23.attn.qkv.bias": "model-00001-of-00004.safetensors",
540
- "visual.blocks.23.attn.qkv.weight": "model-00001-of-00004.safetensors",
541
- "visual.blocks.23.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
542
- "visual.blocks.23.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
543
- "visual.blocks.23.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
544
- "visual.blocks.23.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
545
- "visual.blocks.23.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
546
- "visual.blocks.23.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
547
- "visual.blocks.23.norm1.weight": "model-00001-of-00004.safetensors",
548
- "visual.blocks.23.norm2.weight": "model-00001-of-00004.safetensors",
549
- "visual.blocks.24.attn.proj.bias": "model-00001-of-00004.safetensors",
550
- "visual.blocks.24.attn.proj.weight": "model-00001-of-00004.safetensors",
551
- "visual.blocks.24.attn.qkv.bias": "model-00001-of-00004.safetensors",
552
- "visual.blocks.24.attn.qkv.weight": "model-00001-of-00004.safetensors",
553
- "visual.blocks.24.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
554
- "visual.blocks.24.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
555
- "visual.blocks.24.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
556
- "visual.blocks.24.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
557
- "visual.blocks.24.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
558
- "visual.blocks.24.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
559
- "visual.blocks.24.norm1.weight": "model-00001-of-00004.safetensors",
560
- "visual.blocks.24.norm2.weight": "model-00001-of-00004.safetensors",
561
- "visual.blocks.25.attn.proj.bias": "model-00001-of-00004.safetensors",
562
- "visual.blocks.25.attn.proj.weight": "model-00001-of-00004.safetensors",
563
- "visual.blocks.25.attn.qkv.bias": "model-00001-of-00004.safetensors",
564
- "visual.blocks.25.attn.qkv.weight": "model-00001-of-00004.safetensors",
565
- "visual.blocks.25.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
566
- "visual.blocks.25.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
567
- "visual.blocks.25.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
568
- "visual.blocks.25.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
569
- "visual.blocks.25.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
570
- "visual.blocks.25.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
571
- "visual.blocks.25.norm1.weight": "model-00001-of-00004.safetensors",
572
- "visual.blocks.25.norm2.weight": "model-00001-of-00004.safetensors",
573
- "visual.blocks.26.attn.proj.bias": "model-00001-of-00004.safetensors",
574
- "visual.blocks.26.attn.proj.weight": "model-00001-of-00004.safetensors",
575
- "visual.blocks.26.attn.qkv.bias": "model-00001-of-00004.safetensors",
576
- "visual.blocks.26.attn.qkv.weight": "model-00001-of-00004.safetensors",
577
- "visual.blocks.26.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
578
- "visual.blocks.26.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
579
- "visual.blocks.26.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
580
- "visual.blocks.26.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
581
- "visual.blocks.26.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
582
- "visual.blocks.26.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
583
- "visual.blocks.26.norm1.weight": "model-00001-of-00004.safetensors",
584
- "visual.blocks.26.norm2.weight": "model-00001-of-00004.safetensors",
585
- "visual.blocks.27.attn.proj.bias": "model-00001-of-00004.safetensors",
586
- "visual.blocks.27.attn.proj.weight": "model-00001-of-00004.safetensors",
587
- "visual.blocks.27.attn.qkv.bias": "model-00001-of-00004.safetensors",
588
- "visual.blocks.27.attn.qkv.weight": "model-00001-of-00004.safetensors",
589
- "visual.blocks.27.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
590
- "visual.blocks.27.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
591
- "visual.blocks.27.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
592
- "visual.blocks.27.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
593
- "visual.blocks.27.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
594
- "visual.blocks.27.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
595
- "visual.blocks.27.norm1.weight": "model-00001-of-00004.safetensors",
596
- "visual.blocks.27.norm2.weight": "model-00001-of-00004.safetensors",
597
- "visual.blocks.28.attn.proj.bias": "model-00001-of-00004.safetensors",
598
- "visual.blocks.28.attn.proj.weight": "model-00001-of-00004.safetensors",
599
- "visual.blocks.28.attn.qkv.bias": "model-00001-of-00004.safetensors",
600
- "visual.blocks.28.attn.qkv.weight": "model-00001-of-00004.safetensors",
601
- "visual.blocks.28.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
602
- "visual.blocks.28.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
603
- "visual.blocks.28.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
604
- "visual.blocks.28.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
605
- "visual.blocks.28.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
606
- "visual.blocks.28.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
607
- "visual.blocks.28.norm1.weight": "model-00001-of-00004.safetensors",
608
- "visual.blocks.28.norm2.weight": "model-00001-of-00004.safetensors",
609
- "visual.blocks.29.attn.proj.bias": "model-00001-of-00004.safetensors",
610
- "visual.blocks.29.attn.proj.weight": "model-00001-of-00004.safetensors",
611
- "visual.blocks.29.attn.qkv.bias": "model-00001-of-00004.safetensors",
612
- "visual.blocks.29.attn.qkv.weight": "model-00001-of-00004.safetensors",
613
- "visual.blocks.29.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
614
- "visual.blocks.29.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
615
- "visual.blocks.29.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
616
- "visual.blocks.29.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
617
- "visual.blocks.29.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
618
- "visual.blocks.29.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
619
- "visual.blocks.29.norm1.weight": "model-00001-of-00004.safetensors",
620
- "visual.blocks.29.norm2.weight": "model-00001-of-00004.safetensors",
621
- "visual.blocks.3.attn.proj.bias": "model-00001-of-00004.safetensors",
622
- "visual.blocks.3.attn.proj.weight": "model-00001-of-00004.safetensors",
623
- "visual.blocks.3.attn.qkv.bias": "model-00001-of-00004.safetensors",
624
- "visual.blocks.3.attn.qkv.weight": "model-00001-of-00004.safetensors",
625
- "visual.blocks.3.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
626
- "visual.blocks.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
627
- "visual.blocks.3.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
628
- "visual.blocks.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
629
- "visual.blocks.3.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
630
- "visual.blocks.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
631
- "visual.blocks.3.norm1.weight": "model-00001-of-00004.safetensors",
632
- "visual.blocks.3.norm2.weight": "model-00001-of-00004.safetensors",
633
- "visual.blocks.30.attn.proj.bias": "model-00001-of-00004.safetensors",
634
- "visual.blocks.30.attn.proj.weight": "model-00001-of-00004.safetensors",
635
- "visual.blocks.30.attn.qkv.bias": "model-00001-of-00004.safetensors",
636
- "visual.blocks.30.attn.qkv.weight": "model-00001-of-00004.safetensors",
637
- "visual.blocks.30.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
638
- "visual.blocks.30.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
639
- "visual.blocks.30.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
640
- "visual.blocks.30.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
641
- "visual.blocks.30.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
642
- "visual.blocks.30.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
643
- "visual.blocks.30.norm1.weight": "model-00001-of-00004.safetensors",
644
- "visual.blocks.30.norm2.weight": "model-00001-of-00004.safetensors",
645
- "visual.blocks.31.attn.proj.bias": "model-00001-of-00004.safetensors",
646
- "visual.blocks.31.attn.proj.weight": "model-00001-of-00004.safetensors",
647
- "visual.blocks.31.attn.qkv.bias": "model-00001-of-00004.safetensors",
648
- "visual.blocks.31.attn.qkv.weight": "model-00001-of-00004.safetensors",
649
- "visual.blocks.31.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
650
- "visual.blocks.31.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
651
- "visual.blocks.31.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
652
- "visual.blocks.31.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
653
- "visual.blocks.31.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
654
- "visual.blocks.31.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
655
- "visual.blocks.31.norm1.weight": "model-00001-of-00004.safetensors",
656
- "visual.blocks.31.norm2.weight": "model-00001-of-00004.safetensors",
657
- "visual.blocks.4.attn.proj.bias": "model-00001-of-00004.safetensors",
658
- "visual.blocks.4.attn.proj.weight": "model-00001-of-00004.safetensors",
659
- "visual.blocks.4.attn.qkv.bias": "model-00001-of-00004.safetensors",
660
- "visual.blocks.4.attn.qkv.weight": "model-00001-of-00004.safetensors",
661
- "visual.blocks.4.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
662
- "visual.blocks.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
663
- "visual.blocks.4.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
664
- "visual.blocks.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
665
- "visual.blocks.4.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
666
- "visual.blocks.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
667
- "visual.blocks.4.norm1.weight": "model-00001-of-00004.safetensors",
668
- "visual.blocks.4.norm2.weight": "model-00001-of-00004.safetensors",
669
- "visual.blocks.5.attn.proj.bias": "model-00001-of-00004.safetensors",
670
- "visual.blocks.5.attn.proj.weight": "model-00001-of-00004.safetensors",
671
- "visual.blocks.5.attn.qkv.bias": "model-00001-of-00004.safetensors",
672
- "visual.blocks.5.attn.qkv.weight": "model-00001-of-00004.safetensors",
673
- "visual.blocks.5.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
674
- "visual.blocks.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
675
- "visual.blocks.5.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
676
- "visual.blocks.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
677
- "visual.blocks.5.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
678
- "visual.blocks.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
679
- "visual.blocks.5.norm1.weight": "model-00001-of-00004.safetensors",
680
- "visual.blocks.5.norm2.weight": "model-00001-of-00004.safetensors",
681
- "visual.blocks.6.attn.proj.bias": "model-00001-of-00004.safetensors",
682
- "visual.blocks.6.attn.proj.weight": "model-00001-of-00004.safetensors",
683
- "visual.blocks.6.attn.qkv.bias": "model-00001-of-00004.safetensors",
684
- "visual.blocks.6.attn.qkv.weight": "model-00001-of-00004.safetensors",
685
- "visual.blocks.6.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
686
- "visual.blocks.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
687
- "visual.blocks.6.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
688
- "visual.blocks.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
689
- "visual.blocks.6.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
690
- "visual.blocks.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
691
- "visual.blocks.6.norm1.weight": "model-00001-of-00004.safetensors",
692
- "visual.blocks.6.norm2.weight": "model-00001-of-00004.safetensors",
693
- "visual.blocks.7.attn.proj.bias": "model-00001-of-00004.safetensors",
694
- "visual.blocks.7.attn.proj.weight": "model-00001-of-00004.safetensors",
695
- "visual.blocks.7.attn.qkv.bias": "model-00001-of-00004.safetensors",
696
- "visual.blocks.7.attn.qkv.weight": "model-00001-of-00004.safetensors",
697
- "visual.blocks.7.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
698
- "visual.blocks.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
699
- "visual.blocks.7.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
700
- "visual.blocks.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
701
- "visual.blocks.7.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
702
- "visual.blocks.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
703
- "visual.blocks.7.norm1.weight": "model-00001-of-00004.safetensors",
704
- "visual.blocks.7.norm2.weight": "model-00001-of-00004.safetensors",
705
- "visual.blocks.8.attn.proj.bias": "model-00001-of-00004.safetensors",
706
- "visual.blocks.8.attn.proj.weight": "model-00001-of-00004.safetensors",
707
- "visual.blocks.8.attn.qkv.bias": "model-00001-of-00004.safetensors",
708
- "visual.blocks.8.attn.qkv.weight": "model-00001-of-00004.safetensors",
709
- "visual.blocks.8.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
710
- "visual.blocks.8.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
711
- "visual.blocks.8.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
712
- "visual.blocks.8.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
713
- "visual.blocks.8.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
714
- "visual.blocks.8.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
715
- "visual.blocks.8.norm1.weight": "model-00001-of-00004.safetensors",
716
- "visual.blocks.8.norm2.weight": "model-00001-of-00004.safetensors",
717
- "visual.blocks.9.attn.proj.bias": "model-00001-of-00004.safetensors",
718
- "visual.blocks.9.attn.proj.weight": "model-00001-of-00004.safetensors",
719
- "visual.blocks.9.attn.qkv.bias": "model-00001-of-00004.safetensors",
720
- "visual.blocks.9.attn.qkv.weight": "model-00001-of-00004.safetensors",
721
- "visual.blocks.9.mlp.down_proj.bias": "model-00001-of-00004.safetensors",
722
- "visual.blocks.9.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
723
- "visual.blocks.9.mlp.gate_proj.bias": "model-00001-of-00004.safetensors",
724
- "visual.blocks.9.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
725
- "visual.blocks.9.mlp.up_proj.bias": "model-00001-of-00004.safetensors",
726
- "visual.blocks.9.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
727
- "visual.blocks.9.norm1.weight": "model-00001-of-00004.safetensors",
728
- "visual.blocks.9.norm2.weight": "model-00001-of-00004.safetensors",
729
- "visual.merger.ln_q.weight": "model-00001-of-00004.safetensors",
730
- "visual.merger.mlp.0.bias": "model-00001-of-00004.safetensors",
731
- "visual.merger.mlp.0.weight": "model-00001-of-00004.safetensors",
732
- "visual.merger.mlp.2.bias": "model-00001-of-00004.safetensors",
733
- "visual.merger.mlp.2.weight": "model-00001-of-00004.safetensors",
734
- "visual.patch_embed.proj.weight": "model-00001-of-00004.safetensors"
735
- }
736
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/preprocessor_config.json DELETED
@@ -1,29 +0,0 @@
1
- {
2
- "do_convert_rgb": true,
3
- "do_normalize": true,
4
- "do_rescale": true,
5
- "do_resize": true,
6
- "image_mean": [
7
- 0.48145466,
8
- 0.4578275,
9
- 0.40821073
10
- ],
11
- "image_processor_type": "Qwen2VLImageProcessor",
12
- "image_std": [
13
- 0.26862954,
14
- 0.26130258,
15
- 0.27577711
16
- ],
17
- "max_pixels": 12845056,
18
- "merge_size": 2,
19
- "min_pixels": 3136,
20
- "patch_size": 14,
21
- "processor_class": "Qwen2_5_VLProcessor",
22
- "resample": 3,
23
- "rescale_factor": 0.00392156862745098,
24
- "size": {
25
- "longest_edge": 12845056,
26
- "shortest_edge": 3136
27
- },
28
- "temporal_patch_size": 2
29
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/special_tokens_map.json DELETED
@@ -1,31 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "<|im_start|>",
4
- "<|im_end|>",
5
- "<|object_ref_start|>",
6
- "<|object_ref_end|>",
7
- "<|box_start|>",
8
- "<|box_end|>",
9
- "<|quad_start|>",
10
- "<|quad_end|>",
11
- "<|vision_start|>",
12
- "<|vision_end|>",
13
- "<|vision_pad|>",
14
- "<|image_pad|>",
15
- "<|video_pad|>"
16
- ],
17
- "eos_token": {
18
- "content": "<|im_end|>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- },
24
- "pad_token": {
25
- "content": "<|endoftext|>",
26
- "lstrip": false,
27
- "normalized": false,
28
- "rstrip": false,
29
- "single_word": false
30
- }
31
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba0c439f7be467bf47d12a7e6f9adc6116201056fc60c67f431c679b7c16afc8
3
- size 11422064
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/tokenizer_config.json DELETED
@@ -1,209 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "added_tokens_decoder": {
5
- "151643": {
6
- "content": "<|endoftext|>",
7
- "lstrip": false,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "151644": {
14
- "content": "<|im_start|>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": true
20
- },
21
- "151645": {
22
- "content": "<|im_end|>",
23
- "lstrip": false,
24
- "normalized": false,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": true
28
- },
29
- "151646": {
30
- "content": "<|object_ref_start|>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "151647": {
38
- "content": "<|object_ref_end|>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "151648": {
46
- "content": "<|box_start|>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
- },
53
- "151649": {
54
- "content": "<|box_end|>",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": true
60
- },
61
- "151650": {
62
- "content": "<|quad_start|>",
63
- "lstrip": false,
64
- "normalized": false,
65
- "rstrip": false,
66
- "single_word": false,
67
- "special": true
68
- },
69
- "151651": {
70
- "content": "<|quad_end|>",
71
- "lstrip": false,
72
- "normalized": false,
73
- "rstrip": false,
74
- "single_word": false,
75
- "special": true
76
- },
77
- "151652": {
78
- "content": "<|vision_start|>",
79
- "lstrip": false,
80
- "normalized": false,
81
- "rstrip": false,
82
- "single_word": false,
83
- "special": true
84
- },
85
- "151653": {
86
- "content": "<|vision_end|>",
87
- "lstrip": false,
88
- "normalized": false,
89
- "rstrip": false,
90
- "single_word": false,
91
- "special": true
92
- },
93
- "151654": {
94
- "content": "<|vision_pad|>",
95
- "lstrip": false,
96
- "normalized": false,
97
- "rstrip": false,
98
- "single_word": false,
99
- "special": true
100
- },
101
- "151655": {
102
- "content": "<|image_pad|>",
103
- "lstrip": false,
104
- "normalized": false,
105
- "rstrip": false,
106
- "single_word": false,
107
- "special": true
108
- },
109
- "151656": {
110
- "content": "<|video_pad|>",
111
- "lstrip": false,
112
- "normalized": false,
113
- "rstrip": false,
114
- "single_word": false,
115
- "special": true
116
- },
117
- "151657": {
118
- "content": "<tool_call>",
119
- "lstrip": false,
120
- "normalized": false,
121
- "rstrip": false,
122
- "single_word": false,
123
- "special": false
124
- },
125
- "151658": {
126
- "content": "</tool_call>",
127
- "lstrip": false,
128
- "normalized": false,
129
- "rstrip": false,
130
- "single_word": false,
131
- "special": false
132
- },
133
- "151659": {
134
- "content": "<|fim_prefix|>",
135
- "lstrip": false,
136
- "normalized": false,
137
- "rstrip": false,
138
- "single_word": false,
139
- "special": false
140
- },
141
- "151660": {
142
- "content": "<|fim_middle|>",
143
- "lstrip": false,
144
- "normalized": false,
145
- "rstrip": false,
146
- "single_word": false,
147
- "special": false
148
- },
149
- "151661": {
150
- "content": "<|fim_suffix|>",
151
- "lstrip": false,
152
- "normalized": false,
153
- "rstrip": false,
154
- "single_word": false,
155
- "special": false
156
- },
157
- "151662": {
158
- "content": "<|fim_pad|>",
159
- "lstrip": false,
160
- "normalized": false,
161
- "rstrip": false,
162
- "single_word": false,
163
- "special": false
164
- },
165
- "151663": {
166
- "content": "<|repo_name|>",
167
- "lstrip": false,
168
- "normalized": false,
169
- "rstrip": false,
170
- "single_word": false,
171
- "special": false
172
- },
173
- "151664": {
174
- "content": "<|file_sep|>",
175
- "lstrip": false,
176
- "normalized": false,
177
- "rstrip": false,
178
- "single_word": false,
179
- "special": false
180
- }
181
- },
182
- "additional_special_tokens": [
183
- "<|im_start|>",
184
- "<|im_end|>",
185
- "<|object_ref_start|>",
186
- "<|object_ref_end|>",
187
- "<|box_start|>",
188
- "<|box_end|>",
189
- "<|quad_start|>",
190
- "<|quad_end|>",
191
- "<|vision_start|>",
192
- "<|vision_end|>",
193
- "<|vision_pad|>",
194
- "<|image_pad|>",
195
- "<|video_pad|>"
196
- ],
197
- "bos_token": null,
198
- "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}",
199
- "clean_up_tokenization_spaces": false,
200
- "eos_token": "<|im_end|>",
201
- "errors": "replace",
202
- "extra_special_tokens": {},
203
- "model_max_length": 131072,
204
- "pad_token": "<|endoftext|>",
205
- "processor_class": "Qwen2_5_VLProcessor",
206
- "split_special_tokens": false,
207
- "tokenizer_class": "Qwen2Tokenizer",
208
- "unk_token": null
209
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/trainer_state.json DELETED
@@ -1,3033 +0,0 @@
1
- {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 20.0,
5
- "eval_steps": 500,
6
- "global_step": 2500,
7
- "is_hyper_param_search": false,
8
- "is_local_process_zero": true,
9
- "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 0,
13
- "step": 0,
14
- "train/loss": 2.271090269088745
15
- },
16
- {
17
- "epoch": 0.08,
18
- "grad_norm": 10.1875,
19
- "learning_rate": 9.960000000000001e-06,
20
- "loss": 0.6981,
21
- "step": 10
22
- },
23
- {
24
- "epoch": 0.08,
25
- "step": 10,
26
- "train/loss": 0.17855204641819
27
- },
28
- {
29
- "epoch": 0.16,
30
- "grad_norm": 4.125,
31
- "learning_rate": 9.920000000000002e-06,
32
- "loss": 0.1739,
33
- "step": 20
34
- },
35
- {
36
- "epoch": 0.16,
37
- "step": 20,
38
- "train/loss": 0.19313403964042664
39
- },
40
- {
41
- "epoch": 0.24,
42
- "grad_norm": 6.71875,
43
- "learning_rate": 9.88e-06,
44
- "loss": 0.1152,
45
- "step": 30
46
- },
47
- {
48
- "epoch": 0.24,
49
- "step": 30,
50
- "train/loss": 0.05154286324977875
51
- },
52
- {
53
- "epoch": 0.32,
54
- "grad_norm": 4.53125,
55
- "learning_rate": 9.84e-06,
56
- "loss": 0.0714,
57
- "step": 40
58
- },
59
- {
60
- "epoch": 0.32,
61
- "step": 40,
62
- "train/loss": 0.08214734494686127
63
- },
64
- {
65
- "epoch": 0.4,
66
- "grad_norm": 9.25,
67
- "learning_rate": 9.800000000000001e-06,
68
- "loss": 0.051,
69
- "step": 50
70
- },
71
- {
72
- "epoch": 0.4,
73
- "step": 50,
74
- "train/loss": 0.16986459493637085
75
- },
76
- {
77
- "epoch": 0.48,
78
- "grad_norm": 3.46875,
79
- "learning_rate": 9.760000000000001e-06,
80
- "loss": 0.0643,
81
- "step": 60
82
- },
83
- {
84
- "epoch": 0.48,
85
- "step": 60,
86
- "train/loss": 0.04494616761803627
87
- },
88
- {
89
- "epoch": 0.56,
90
- "grad_norm": 6.4375,
91
- "learning_rate": 9.72e-06,
92
- "loss": 0.0315,
93
- "step": 70
94
- },
95
- {
96
- "epoch": 0.56,
97
- "step": 70,
98
- "train/loss": 0.008576815016567707
99
- },
100
- {
101
- "epoch": 0.64,
102
- "grad_norm": 12.625,
103
- "learning_rate": 9.68e-06,
104
- "loss": 0.0293,
105
- "step": 80
106
- },
107
- {
108
- "epoch": 0.64,
109
- "step": 80,
110
- "train/loss": 0.05587716028094292
111
- },
112
- {
113
- "epoch": 0.72,
114
- "grad_norm": 4.25,
115
- "learning_rate": 9.640000000000001e-06,
116
- "loss": 0.0236,
117
- "step": 90
118
- },
119
- {
120
- "epoch": 0.72,
121
- "step": 90,
122
- "train/loss": 0.011845740489661694
123
- },
124
- {
125
- "epoch": 0.8,
126
- "grad_norm": 3.4375,
127
- "learning_rate": 9.600000000000001e-06,
128
- "loss": 0.0316,
129
- "step": 100
130
- },
131
- {
132
- "epoch": 0.8,
133
- "step": 100,
134
- "train/loss": 0.048473458737134933
135
- },
136
- {
137
- "epoch": 0.88,
138
- "grad_norm": 2.296875,
139
- "learning_rate": 9.56e-06,
140
- "loss": 0.0218,
141
- "step": 110
142
- },
143
- {
144
- "epoch": 0.88,
145
- "step": 110,
146
- "train/loss": 0.00870122853666544
147
- },
148
- {
149
- "epoch": 0.96,
150
- "grad_norm": 6.84375,
151
- "learning_rate": 9.52e-06,
152
- "loss": 0.0248,
153
- "step": 120
154
- },
155
- {
156
- "epoch": 0.96,
157
- "step": 120,
158
- "train/loss": 0.021771380677819252
159
- },
160
- {
161
- "epoch": 1.04,
162
- "grad_norm": 3.25,
163
- "learning_rate": 9.48e-06,
164
- "loss": 0.019,
165
- "step": 130
166
- },
167
- {
168
- "epoch": 1.04,
169
- "step": 130,
170
- "train/loss": 0.015329583548009396
171
- },
172
- {
173
- "epoch": 1.12,
174
- "grad_norm": 2.46875,
175
- "learning_rate": 9.440000000000001e-06,
176
- "loss": 0.0167,
177
- "step": 140
178
- },
179
- {
180
- "epoch": 1.12,
181
- "step": 140,
182
- "train/loss": 0.03584172576665878
183
- },
184
- {
185
- "epoch": 1.2,
186
- "grad_norm": 3.359375,
187
- "learning_rate": 9.4e-06,
188
- "loss": 0.0271,
189
- "step": 150
190
- },
191
- {
192
- "epoch": 1.2,
193
- "step": 150,
194
- "train/loss": 0.028160186484456062
195
- },
196
- {
197
- "epoch": 1.28,
198
- "grad_norm": 2.390625,
199
- "learning_rate": 9.360000000000002e-06,
200
- "loss": 0.0198,
201
- "step": 160
202
- },
203
- {
204
- "epoch": 1.28,
205
- "step": 160,
206
- "train/loss": 0.009682728908956051
207
- },
208
- {
209
- "epoch": 1.3599999999999999,
210
- "grad_norm": 1.9140625,
211
- "learning_rate": 9.32e-06,
212
- "loss": 0.0156,
213
- "step": 170
214
- },
215
- {
216
- "epoch": 1.3599999999999999,
217
- "step": 170,
218
- "train/loss": 0.001779431477189064
219
- },
220
- {
221
- "epoch": 1.44,
222
- "grad_norm": 9.8125,
223
- "learning_rate": 9.280000000000001e-06,
224
- "loss": 0.0172,
225
- "step": 180
226
- },
227
- {
228
- "epoch": 1.44,
229
- "step": 180,
230
- "train/loss": 0.024483440443873405
231
- },
232
- {
233
- "epoch": 1.52,
234
- "grad_norm": 2.453125,
235
- "learning_rate": 9.240000000000001e-06,
236
- "loss": 0.0119,
237
- "step": 190
238
- },
239
- {
240
- "epoch": 1.52,
241
- "step": 190,
242
- "train/loss": 0.0002734360168687999
243
- },
244
- {
245
- "epoch": 1.6,
246
- "grad_norm": 2.296875,
247
- "learning_rate": 9.200000000000002e-06,
248
- "loss": 0.012,
249
- "step": 200
250
- },
251
- {
252
- "epoch": 1.6,
253
- "step": 200,
254
- "train/loss": 0.008814319968223572
255
- },
256
- {
257
- "epoch": 1.6800000000000002,
258
- "grad_norm": 0.79296875,
259
- "learning_rate": 9.16e-06,
260
- "loss": 0.0108,
261
- "step": 210
262
- },
263
- {
264
- "epoch": 1.6800000000000002,
265
- "step": 210,
266
- "train/loss": 0.004692906513810158
267
- },
268
- {
269
- "epoch": 1.76,
270
- "grad_norm": 1.84375,
271
- "learning_rate": 9.12e-06,
272
- "loss": 0.0134,
273
- "step": 220
274
- },
275
- {
276
- "epoch": 1.76,
277
- "step": 220,
278
- "train/loss": 0.005740709602832794
279
- },
280
- {
281
- "epoch": 1.8399999999999999,
282
- "grad_norm": 3.625,
283
- "learning_rate": 9.080000000000001e-06,
284
- "loss": 0.0072,
285
- "step": 230
286
- },
287
- {
288
- "epoch": 1.8399999999999999,
289
- "step": 230,
290
- "train/loss": 0.0017300192266702652
291
- },
292
- {
293
- "epoch": 1.92,
294
- "grad_norm": 1.5234375,
295
- "learning_rate": 9.040000000000002e-06,
296
- "loss": 0.0068,
297
- "step": 240
298
- },
299
- {
300
- "epoch": 1.92,
301
- "step": 240,
302
- "train/loss": 0.01101005170494318
303
- },
304
- {
305
- "epoch": 2.0,
306
- "grad_norm": 1.4375,
307
- "learning_rate": 9e-06,
308
- "loss": 0.0128,
309
- "step": 250
310
- },
311
- {
312
- "epoch": 2.0,
313
- "step": 250,
314
- "train/loss": 0.003425895469263196
315
- },
316
- {
317
- "epoch": 2.08,
318
- "grad_norm": 2.6875,
319
- "learning_rate": 8.96e-06,
320
- "loss": 0.0074,
321
- "step": 260
322
- },
323
- {
324
- "epoch": 2.08,
325
- "step": 260,
326
- "train/loss": 0.026302414014935493
327
- },
328
- {
329
- "epoch": 2.16,
330
- "grad_norm": 1.2421875,
331
- "learning_rate": 8.920000000000001e-06,
332
- "loss": 0.0102,
333
- "step": 270
334
- },
335
- {
336
- "epoch": 2.16,
337
- "step": 270,
338
- "train/loss": 0.0006555073196068406
339
- },
340
- {
341
- "epoch": 2.24,
342
- "grad_norm": 1.6640625,
343
- "learning_rate": 8.880000000000001e-06,
344
- "loss": 0.009,
345
- "step": 280
346
- },
347
- {
348
- "epoch": 2.24,
349
- "step": 280,
350
- "train/loss": 0.005413789767771959
351
- },
352
- {
353
- "epoch": 2.32,
354
- "grad_norm": 1.171875,
355
- "learning_rate": 8.84e-06,
356
- "loss": 0.0059,
357
- "step": 290
358
- },
359
- {
360
- "epoch": 2.32,
361
- "step": 290,
362
- "train/loss": 0.016869384795427322
363
- },
364
- {
365
- "epoch": 2.4,
366
- "grad_norm": 2.5625,
367
- "learning_rate": 8.8e-06,
368
- "loss": 0.0148,
369
- "step": 300
370
- },
371
- {
372
- "epoch": 2.4,
373
- "step": 300,
374
- "train/loss": 5.8237197663402185e-05
375
- },
376
- {
377
- "epoch": 2.48,
378
- "grad_norm": 1.296875,
379
- "learning_rate": 8.76e-06,
380
- "loss": 0.0109,
381
- "step": 310
382
- },
383
- {
384
- "epoch": 2.48,
385
- "step": 310,
386
- "train/loss": 0.0001535036863060668
387
- },
388
- {
389
- "epoch": 2.56,
390
- "grad_norm": 3.875,
391
- "learning_rate": 8.720000000000001e-06,
392
- "loss": 0.0042,
393
- "step": 320
394
- },
395
- {
396
- "epoch": 2.56,
397
- "step": 320,
398
- "train/loss": 0.0072115445509552956
399
- },
400
- {
401
- "epoch": 2.64,
402
- "grad_norm": 1.015625,
403
- "learning_rate": 8.68e-06,
404
- "loss": 0.0068,
405
- "step": 330
406
- },
407
- {
408
- "epoch": 2.64,
409
- "step": 330,
410
- "train/loss": 0.004820008762180805
411
- },
412
- {
413
- "epoch": 2.7199999999999998,
414
- "grad_norm": 6.34375,
415
- "learning_rate": 8.64e-06,
416
- "loss": 0.0132,
417
- "step": 340
418
- },
419
- {
420
- "epoch": 2.7199999999999998,
421
- "step": 340,
422
- "train/loss": 0.003863546997308731
423
- },
424
- {
425
- "epoch": 2.8,
426
- "grad_norm": 1.65625,
427
- "learning_rate": 8.6e-06,
428
- "loss": 0.0037,
429
- "step": 350
430
- },
431
- {
432
- "epoch": 2.8,
433
- "step": 350,
434
- "train/loss": 0.000956513627897948
435
- },
436
- {
437
- "epoch": 2.88,
438
- "grad_norm": 0.224609375,
439
- "learning_rate": 8.560000000000001e-06,
440
- "loss": 0.0018,
441
- "step": 360
442
- },
443
- {
444
- "epoch": 2.88,
445
- "step": 360,
446
- "train/loss": 0.0001041244322550483
447
- },
448
- {
449
- "epoch": 2.96,
450
- "grad_norm": 1.2734375,
451
- "learning_rate": 8.52e-06,
452
- "loss": 0.0089,
453
- "step": 370
454
- },
455
- {
456
- "epoch": 2.96,
457
- "step": 370,
458
- "train/loss": 0.014187252148985863
459
- },
460
- {
461
- "epoch": 3.04,
462
- "grad_norm": 0.76953125,
463
- "learning_rate": 8.48e-06,
464
- "loss": 0.0053,
465
- "step": 380
466
- },
467
- {
468
- "epoch": 3.04,
469
- "step": 380,
470
- "train/loss": 0.00017445831326767802
471
- },
472
- {
473
- "epoch": 3.12,
474
- "grad_norm": 0.4140625,
475
- "learning_rate": 8.44e-06,
476
- "loss": 0.0038,
477
- "step": 390
478
- },
479
- {
480
- "epoch": 3.12,
481
- "step": 390,
482
- "train/loss": 0.0015271755401045084
483
- },
484
- {
485
- "epoch": 3.2,
486
- "grad_norm": 1.2421875,
487
- "learning_rate": 8.400000000000001e-06,
488
- "loss": 0.0046,
489
- "step": 400
490
- },
491
- {
492
- "epoch": 3.2,
493
- "step": 400,
494
- "train/loss": 0.0023774108849465847
495
- },
496
- {
497
- "epoch": 3.2800000000000002,
498
- "grad_norm": 0.54296875,
499
- "learning_rate": 8.36e-06,
500
- "loss": 0.0053,
501
- "step": 410
502
- },
503
- {
504
- "epoch": 3.2800000000000002,
505
- "step": 410,
506
- "train/loss": 0.0044172233901917934
507
- },
508
- {
509
- "epoch": 3.36,
510
- "grad_norm": 0.365234375,
511
- "learning_rate": 8.32e-06,
512
- "loss": 0.0045,
513
- "step": 420
514
- },
515
- {
516
- "epoch": 3.36,
517
- "step": 420,
518
- "train/loss": 0.0023545583244413137
519
- },
520
- {
521
- "epoch": 3.44,
522
- "grad_norm": 0.28515625,
523
- "learning_rate": 8.28e-06,
524
- "loss": 0.0043,
525
- "step": 430
526
- },
527
- {
528
- "epoch": 3.44,
529
- "step": 430,
530
- "train/loss": 3.492781615932472e-05
531
- },
532
- {
533
- "epoch": 3.52,
534
- "grad_norm": 0.95703125,
535
- "learning_rate": 8.24e-06,
536
- "loss": 0.002,
537
- "step": 440
538
- },
539
- {
540
- "epoch": 3.52,
541
- "step": 440,
542
- "train/loss": 1.694618731562514e-05
543
- },
544
- {
545
- "epoch": 3.6,
546
- "grad_norm": 0.5078125,
547
- "learning_rate": 8.2e-06,
548
- "loss": 0.0025,
549
- "step": 450
550
- },
551
- {
552
- "epoch": 3.6,
553
- "step": 450,
554
- "train/loss": 1.0210129403276369e-05
555
- },
556
- {
557
- "epoch": 3.68,
558
- "grad_norm": 1.6640625,
559
- "learning_rate": 8.16e-06,
560
- "loss": 0.0032,
561
- "step": 460
562
- },
563
- {
564
- "epoch": 3.68,
565
- "step": 460,
566
- "train/loss": 4.278627966414206e-05
567
- },
568
- {
569
- "epoch": 3.76,
570
- "grad_norm": 0.0966796875,
571
- "learning_rate": 8.120000000000002e-06,
572
- "loss": 0.0019,
573
- "step": 470
574
- },
575
- {
576
- "epoch": 3.76,
577
- "step": 470,
578
- "train/loss": 1.557427276566159e-05
579
- },
580
- {
581
- "epoch": 3.84,
582
- "grad_norm": 2.078125,
583
- "learning_rate": 8.08e-06,
584
- "loss": 0.0025,
585
- "step": 480
586
- },
587
- {
588
- "epoch": 3.84,
589
- "step": 480,
590
- "train/loss": 0.0007325322949327528
591
- },
592
- {
593
- "epoch": 3.92,
594
- "grad_norm": 0.126953125,
595
- "learning_rate": 8.040000000000001e-06,
596
- "loss": 0.0028,
597
- "step": 490
598
- },
599
- {
600
- "epoch": 3.92,
601
- "step": 490,
602
- "train/loss": 4.8548095946898684e-05
603
- },
604
- {
605
- "epoch": 4.0,
606
- "grad_norm": 0.59765625,
607
- "learning_rate": 8.000000000000001e-06,
608
- "loss": 0.0028,
609
- "step": 500
610
- },
611
- {
612
- "epoch": 4.0,
613
- "step": 500,
614
- "train/loss": 0.0001273242523893714
615
- },
616
- {
617
- "epoch": 4.08,
618
- "grad_norm": 0.8046875,
619
- "learning_rate": 7.960000000000002e-06,
620
- "loss": 0.0009,
621
- "step": 510
622
- },
623
- {
624
- "epoch": 4.08,
625
- "step": 510,
626
- "train/loss": 0.004483949858695269
627
- },
628
- {
629
- "epoch": 4.16,
630
- "grad_norm": 0.3125,
631
- "learning_rate": 7.92e-06,
632
- "loss": 0.0011,
633
- "step": 520
634
- },
635
- {
636
- "epoch": 4.16,
637
- "step": 520,
638
- "train/loss": 0.002184318844228983
639
- },
640
- {
641
- "epoch": 4.24,
642
- "grad_norm": 1.8671875,
643
- "learning_rate": 7.88e-06,
644
- "loss": 0.0016,
645
- "step": 530
646
- },
647
- {
648
- "epoch": 4.24,
649
- "step": 530,
650
- "train/loss": 5.333217814040836e-06
651
- },
652
- {
653
- "epoch": 4.32,
654
- "grad_norm": 0.7734375,
655
- "learning_rate": 7.840000000000001e-06,
656
- "loss": 0.0013,
657
- "step": 540
658
- },
659
- {
660
- "epoch": 4.32,
661
- "step": 540,
662
- "train/loss": 2.3022841560305096e-05
663
- },
664
- {
665
- "epoch": 4.4,
666
- "grad_norm": 0.0235595703125,
667
- "learning_rate": 7.800000000000002e-06,
668
- "loss": 0.0005,
669
- "step": 550
670
- },
671
- {
672
- "epoch": 4.4,
673
- "step": 550,
674
- "train/loss": 1.1525683476065751e-05
675
- },
676
- {
677
- "epoch": 4.48,
678
- "grad_norm": 0.09033203125,
679
- "learning_rate": 7.76e-06,
680
- "loss": 0.0009,
681
- "step": 560
682
- },
683
- {
684
- "epoch": 4.48,
685
- "step": 560,
686
- "train/loss": 0.0002542025758884847
687
- },
688
- {
689
- "epoch": 4.5600000000000005,
690
- "grad_norm": 0.02392578125,
691
- "learning_rate": 7.72e-06,
692
- "loss": 0.0009,
693
- "step": 570
694
- },
695
- {
696
- "epoch": 4.5600000000000005,
697
- "step": 570,
698
- "train/loss": 8.229686500271782e-06
699
- },
700
- {
701
- "epoch": 4.64,
702
- "grad_norm": 0.0164794921875,
703
- "learning_rate": 7.680000000000001e-06,
704
- "loss": 0.0006,
705
- "step": 580
706
- },
707
- {
708
- "epoch": 4.64,
709
- "step": 580,
710
- "train/loss": 0.0002689993707463145
711
- },
712
- {
713
- "epoch": 4.72,
714
- "grad_norm": 0.0194091796875,
715
- "learning_rate": 7.640000000000001e-06,
716
- "loss": 0.0002,
717
- "step": 590
718
- },
719
- {
720
- "epoch": 4.72,
721
- "step": 590,
722
- "train/loss": 6.554759693244705e-06
723
- },
724
- {
725
- "epoch": 4.8,
726
- "grad_norm": 0.0079345703125,
727
- "learning_rate": 7.600000000000001e-06,
728
- "loss": 0.0012,
729
- "step": 600
730
- },
731
- {
732
- "epoch": 4.8,
733
- "step": 600,
734
- "train/loss": 3.557118907338008e-05
735
- },
736
- {
737
- "epoch": 4.88,
738
- "grad_norm": 0.90234375,
739
- "learning_rate": 7.5600000000000005e-06,
740
- "loss": 0.0006,
741
- "step": 610
742
- },
743
- {
744
- "epoch": 4.88,
745
- "step": 610,
746
- "train/loss": 0.00013326486805453897
747
- },
748
- {
749
- "epoch": 4.96,
750
- "grad_norm": 0.1943359375,
751
- "learning_rate": 7.520000000000001e-06,
752
- "loss": 0.0013,
753
- "step": 620
754
- },
755
- {
756
- "epoch": 4.96,
757
- "step": 620,
758
- "train/loss": 0.00026195202372036874
759
- },
760
- {
761
- "epoch": 5.04,
762
- "grad_norm": 0.2109375,
763
- "learning_rate": 7.48e-06,
764
- "loss": 0.0007,
765
- "step": 630
766
- },
767
- {
768
- "epoch": 5.04,
769
- "step": 630,
770
- "train/loss": 4.716463081422262e-06
771
- },
772
- {
773
- "epoch": 5.12,
774
- "grad_norm": 0.73046875,
775
- "learning_rate": 7.440000000000001e-06,
776
- "loss": 0.001,
777
- "step": 640
778
- },
779
- {
780
- "epoch": 5.12,
781
- "step": 640,
782
- "train/loss": 0.0009888594504445791
783
- },
784
- {
785
- "epoch": 5.2,
786
- "grad_norm": 0.38671875,
787
- "learning_rate": 7.4e-06,
788
- "loss": 0.0006,
789
- "step": 650
790
- },
791
- {
792
- "epoch": 5.2,
793
- "step": 650,
794
- "train/loss": 2.9275020096974913e-06
795
- },
796
- {
797
- "epoch": 5.28,
798
- "grad_norm": 0.9921875,
799
- "learning_rate": 7.360000000000001e-06,
800
- "loss": 0.0013,
801
- "step": 660
802
- },
803
- {
804
- "epoch": 5.28,
805
- "step": 660,
806
- "train/loss": 0.0005347562255337834
807
- },
808
- {
809
- "epoch": 5.36,
810
- "grad_norm": 0.27734375,
811
- "learning_rate": 7.32e-06,
812
- "loss": 0.0006,
813
- "step": 670
814
- },
815
- {
816
- "epoch": 5.36,
817
- "step": 670,
818
- "train/loss": 5.418706132331863e-05
819
- },
820
- {
821
- "epoch": 5.44,
822
- "grad_norm": 0.00543212890625,
823
- "learning_rate": 7.280000000000001e-06,
824
- "loss": 0.0005,
825
- "step": 680
826
- },
827
- {
828
- "epoch": 5.44,
829
- "step": 680,
830
- "train/loss": 0.00025511899730190635
831
- },
832
- {
833
- "epoch": 5.52,
834
- "grad_norm": 0.451171875,
835
- "learning_rate": 7.24e-06,
836
- "loss": 0.0003,
837
- "step": 690
838
- },
839
- {
840
- "epoch": 5.52,
841
- "step": 690,
842
- "train/loss": 6.210683750396129e-06
843
- },
844
- {
845
- "epoch": 5.6,
846
- "grad_norm": 0.21875,
847
- "learning_rate": 7.2000000000000005e-06,
848
- "loss": 0.0007,
849
- "step": 700
850
- },
851
- {
852
- "epoch": 5.6,
853
- "step": 700,
854
- "train/loss": 0.0006609788397327065
855
- },
856
- {
857
- "epoch": 5.68,
858
- "grad_norm": 0.0020599365234375,
859
- "learning_rate": 7.16e-06,
860
- "loss": 0.0005,
861
- "step": 710
862
- },
863
- {
864
- "epoch": 5.68,
865
- "step": 710,
866
- "train/loss": 2.6659903596737422e-05
867
- },
868
- {
869
- "epoch": 5.76,
870
- "grad_norm": 0.224609375,
871
- "learning_rate": 7.1200000000000004e-06,
872
- "loss": 0.0002,
873
- "step": 720
874
- },
875
- {
876
- "epoch": 5.76,
877
- "step": 720,
878
- "train/loss": 7.112800813047215e-05
879
- },
880
- {
881
- "epoch": 5.84,
882
- "grad_norm": 0.490234375,
883
- "learning_rate": 7.08e-06,
884
- "loss": 0.0005,
885
- "step": 730
886
- },
887
- {
888
- "epoch": 5.84,
889
- "step": 730,
890
- "train/loss": 8.572215847379994e-06
891
- },
892
- {
893
- "epoch": 5.92,
894
- "grad_norm": 0.00921630859375,
895
- "learning_rate": 7.04e-06,
896
- "loss": 0.0002,
897
- "step": 740
898
- },
899
- {
900
- "epoch": 5.92,
901
- "step": 740,
902
- "train/loss": 0.0024293933529406786
903
- },
904
- {
905
- "epoch": 6.0,
906
- "grad_norm": 0.0693359375,
907
- "learning_rate": 7e-06,
908
- "loss": 0.0003,
909
- "step": 750
910
- },
911
- {
912
- "epoch": 6.0,
913
- "step": 750,
914
- "train/loss": 1.0925517017312814e-05
915
- },
916
- {
917
- "epoch": 6.08,
918
- "grad_norm": 0.0079345703125,
919
- "learning_rate": 6.96e-06,
920
- "loss": 0.0002,
921
- "step": 760
922
- },
923
- {
924
- "epoch": 6.08,
925
- "step": 760,
926
- "train/loss": 6.2842104853189085e-06
927
- },
928
- {
929
- "epoch": 6.16,
930
- "grad_norm": 0.091796875,
931
- "learning_rate": 6.92e-06,
932
- "loss": 0.0005,
933
- "step": 770
934
- },
935
- {
936
- "epoch": 6.16,
937
- "step": 770,
938
- "train/loss": 1.9141054508509114e-05
939
- },
940
- {
941
- "epoch": 6.24,
942
- "grad_norm": 0.01165771484375,
943
- "learning_rate": 6.88e-06,
944
- "loss": 0.0002,
945
- "step": 780
946
- },
947
- {
948
- "epoch": 6.24,
949
- "step": 780,
950
- "train/loss": 8.359744242625311e-06
951
- },
952
- {
953
- "epoch": 6.32,
954
- "grad_norm": 0.0216064453125,
955
- "learning_rate": 6.8400000000000014e-06,
956
- "loss": 0.0001,
957
- "step": 790
958
- },
959
- {
960
- "epoch": 6.32,
961
- "step": 790,
962
- "train/loss": 2.0496892830124125e-05
963
- },
964
- {
965
- "epoch": 6.4,
966
- "grad_norm": 0.01904296875,
967
- "learning_rate": 6.800000000000001e-06,
968
- "loss": 0.0001,
969
- "step": 800
970
- },
971
- {
972
- "epoch": 6.4,
973
- "step": 800,
974
- "train/loss": 5.762509317719378e-06
975
- },
976
- {
977
- "epoch": 6.48,
978
- "grad_norm": 0.00165557861328125,
979
- "learning_rate": 6.760000000000001e-06,
980
- "loss": 0.0002,
981
- "step": 810
982
- },
983
- {
984
- "epoch": 6.48,
985
- "step": 810,
986
- "train/loss": 2.0831541405641474e-05
987
- },
988
- {
989
- "epoch": 6.5600000000000005,
990
- "grad_norm": 0.19140625,
991
- "learning_rate": 6.720000000000001e-06,
992
- "loss": 0.001,
993
- "step": 820
994
- },
995
- {
996
- "epoch": 6.5600000000000005,
997
- "step": 820,
998
- "train/loss": 0.0008487328886985779
999
- },
1000
- {
1001
- "epoch": 6.64,
1002
- "grad_norm": 0.01806640625,
1003
- "learning_rate": 6.680000000000001e-06,
1004
- "loss": 0.0,
1005
- "step": 830
1006
- },
1007
- {
1008
- "epoch": 6.64,
1009
- "step": 830,
1010
- "train/loss": 0.0021144477650523186
1011
- },
1012
- {
1013
- "epoch": 6.72,
1014
- "grad_norm": 0.09423828125,
1015
- "learning_rate": 6.640000000000001e-06,
1016
- "loss": 0.0001,
1017
- "step": 840
1018
- },
1019
- {
1020
- "epoch": 6.72,
1021
- "step": 840,
1022
- "train/loss": 1.730973417579662e-05
1023
- },
1024
- {
1025
- "epoch": 6.8,
1026
- "grad_norm": 0.00909423828125,
1027
- "learning_rate": 6.600000000000001e-06,
1028
- "loss": 0.0004,
1029
- "step": 850
1030
- },
1031
- {
1032
- "epoch": 6.8,
1033
- "step": 850,
1034
- "train/loss": 1.7513702914584428e-06
1035
- },
1036
- {
1037
- "epoch": 6.88,
1038
- "grad_norm": 0.00787353515625,
1039
- "learning_rate": 6.560000000000001e-06,
1040
- "loss": 0.0003,
1041
- "step": 860
1042
- },
1043
- {
1044
- "epoch": 6.88,
1045
- "step": 860,
1046
- "train/loss": 0.0036166643258184195
1047
- },
1048
- {
1049
- "epoch": 6.96,
1050
- "grad_norm": 0.00238037109375,
1051
- "learning_rate": 6.520000000000001e-06,
1052
- "loss": 0.0004,
1053
- "step": 870
1054
- },
1055
- {
1056
- "epoch": 6.96,
1057
- "step": 870,
1058
- "train/loss": 1.9699104996107053e-06
1059
- },
1060
- {
1061
- "epoch": 7.04,
1062
- "grad_norm": 0.0185546875,
1063
- "learning_rate": 6.480000000000001e-06,
1064
- "loss": 0.0,
1065
- "step": 880
1066
- },
1067
- {
1068
- "epoch": 7.04,
1069
- "step": 880,
1070
- "train/loss": 4.418452135723783e-06
1071
- },
1072
- {
1073
- "epoch": 7.12,
1074
- "grad_norm": 0.0159912109375,
1075
- "learning_rate": 6.440000000000001e-06,
1076
- "loss": 0.0,
1077
- "step": 890
1078
- },
1079
- {
1080
- "epoch": 7.12,
1081
- "step": 890,
1082
- "train/loss": 7.8619332271046e-06
1083
- },
1084
- {
1085
- "epoch": 7.2,
1086
- "grad_norm": 0.01361083984375,
1087
- "learning_rate": 6.4000000000000006e-06,
1088
- "loss": 0.0001,
1089
- "step": 900
1090
- },
1091
- {
1092
- "epoch": 7.2,
1093
- "step": 900,
1094
- "train/loss": 1.224165316671133e-05
1095
- },
1096
- {
1097
- "epoch": 7.28,
1098
- "grad_norm": 0.2158203125,
1099
- "learning_rate": 6.360000000000001e-06,
1100
- "loss": 0.0001,
1101
- "step": 910
1102
- },
1103
- {
1104
- "epoch": 7.28,
1105
- "step": 910,
1106
- "train/loss": 1.74042031630961e-06
1107
- },
1108
- {
1109
- "epoch": 7.36,
1110
- "grad_norm": 0.00421142578125,
1111
- "learning_rate": 6.3200000000000005e-06,
1112
- "loss": 0.0,
1113
- "step": 920
1114
- },
1115
- {
1116
- "epoch": 7.36,
1117
- "step": 920,
1118
- "train/loss": 1.9013602923223516e-06
1119
- },
1120
- {
1121
- "epoch": 7.44,
1122
- "grad_norm": 0.005889892578125,
1123
- "learning_rate": 6.280000000000001e-06,
1124
- "loss": 0.0002,
1125
- "step": 930
1126
- },
1127
- {
1128
- "epoch": 7.44,
1129
- "step": 930,
1130
- "train/loss": 1.5262668966897763e-05
1131
- },
1132
- {
1133
- "epoch": 7.52,
1134
- "grad_norm": 0.01470947265625,
1135
- "learning_rate": 6.24e-06,
1136
- "loss": 0.0001,
1137
- "step": 940
1138
- },
1139
- {
1140
- "epoch": 7.52,
1141
- "step": 940,
1142
- "train/loss": 1.0542215932218824e-05
1143
- },
1144
- {
1145
- "epoch": 7.6,
1146
- "grad_norm": 0.00921630859375,
1147
- "learning_rate": 6.200000000000001e-06,
1148
- "loss": 0.0001,
1149
- "step": 950
1150
- },
1151
- {
1152
- "epoch": 7.6,
1153
- "step": 950,
1154
- "train/loss": 1.299941595789278e-05
1155
- },
1156
- {
1157
- "epoch": 7.68,
1158
- "grad_norm": 0.0130615234375,
1159
- "learning_rate": 6.16e-06,
1160
- "loss": 0.0003,
1161
- "step": 960
1162
- },
1163
- {
1164
- "epoch": 7.68,
1165
- "step": 960,
1166
- "train/loss": 8.325143426191062e-06
1167
- },
1168
- {
1169
- "epoch": 7.76,
1170
- "grad_norm": 0.08984375,
1171
- "learning_rate": 6.120000000000001e-06,
1172
- "loss": 0.0011,
1173
- "step": 970
1174
- },
1175
- {
1176
- "epoch": 7.76,
1177
- "step": 970,
1178
- "train/loss": 9.192517609335482e-06
1179
- },
1180
- {
1181
- "epoch": 7.84,
1182
- "grad_norm": 0.0029449462890625,
1183
- "learning_rate": 6.08e-06,
1184
- "loss": 0.0001,
1185
- "step": 980
1186
- },
1187
- {
1188
- "epoch": 7.84,
1189
- "step": 980,
1190
- "train/loss": 1.614072607480921e-05
1191
- },
1192
- {
1193
- "epoch": 7.92,
1194
- "grad_norm": 0.0093994140625,
1195
- "learning_rate": 6.040000000000001e-06,
1196
- "loss": 0.0,
1197
- "step": 990
1198
- },
1199
- {
1200
- "epoch": 7.92,
1201
- "step": 990,
1202
- "train/loss": 3.7631343730026856e-05
1203
- },
1204
- {
1205
- "epoch": 8.0,
1206
- "grad_norm": 0.01611328125,
1207
- "learning_rate": 6e-06,
1208
- "loss": 0.0,
1209
- "step": 1000
1210
- },
1211
- {
1212
- "epoch": 8.0,
1213
- "step": 1000,
1214
- "train/loss": 1.4566069694410544e-05
1215
- },
1216
- {
1217
- "epoch": 8.08,
1218
- "grad_norm": 0.0245361328125,
1219
- "learning_rate": 5.9600000000000005e-06,
1220
- "loss": 0.0,
1221
- "step": 1010
1222
- },
1223
- {
1224
- "epoch": 8.08,
1225
- "step": 1010,
1226
- "train/loss": 1.8284432371729054e-05
1227
- },
1228
- {
1229
- "epoch": 8.16,
1230
- "grad_norm": 0.07275390625,
1231
- "learning_rate": 5.92e-06,
1232
- "loss": 0.0,
1233
- "step": 1020
1234
- },
1235
- {
1236
- "epoch": 8.16,
1237
- "step": 1020,
1238
- "train/loss": 1.2149367876190809e-06
1239
- },
1240
- {
1241
- "epoch": 8.24,
1242
- "grad_norm": 0.0021514892578125,
1243
- "learning_rate": 5.8800000000000005e-06,
1244
- "loss": 0.0,
1245
- "step": 1030
1246
- },
1247
- {
1248
- "epoch": 8.24,
1249
- "step": 1030,
1250
- "train/loss": 1.8189823094871826e-05
1251
- },
1252
- {
1253
- "epoch": 8.32,
1254
- "grad_norm": 0.01190185546875,
1255
- "learning_rate": 5.84e-06,
1256
- "loss": 0.0005,
1257
- "step": 1040
1258
- },
1259
- {
1260
- "epoch": 8.32,
1261
- "step": 1040,
1262
- "train/loss": 3.7805109514010837e-06
1263
- },
1264
- {
1265
- "epoch": 8.4,
1266
- "grad_norm": 0.00323486328125,
1267
- "learning_rate": 5.8e-06,
1268
- "loss": 0.0,
1269
- "step": 1050
1270
- },
1271
- {
1272
- "epoch": 8.4,
1273
- "step": 1050,
1274
- "train/loss": 1.2520137715910096e-05
1275
- },
1276
- {
1277
- "epoch": 8.48,
1278
- "grad_norm": 0.0034637451171875,
1279
- "learning_rate": 5.76e-06,
1280
- "loss": 0.0002,
1281
- "step": 1060
1282
- },
1283
- {
1284
- "epoch": 8.48,
1285
- "step": 1060,
1286
- "train/loss": 4.325816007622052e-06
1287
- },
1288
- {
1289
- "epoch": 8.56,
1290
- "grad_norm": 0.03759765625,
1291
- "learning_rate": 5.72e-06,
1292
- "loss": 0.0002,
1293
- "step": 1070
1294
- },
1295
- {
1296
- "epoch": 8.56,
1297
- "step": 1070,
1298
- "train/loss": 1.5546473832728225e-06
1299
- },
1300
- {
1301
- "epoch": 8.64,
1302
- "grad_norm": 0.0108642578125,
1303
- "learning_rate": 5.68e-06,
1304
- "loss": 0.0001,
1305
- "step": 1080
1306
- },
1307
- {
1308
- "epoch": 8.64,
1309
- "step": 1080,
1310
- "train/loss": 8.575192623538896e-06
1311
- },
1312
- {
1313
- "epoch": 8.72,
1314
- "grad_norm": 0.002593994140625,
1315
- "learning_rate": 5.64e-06,
1316
- "loss": 0.0,
1317
- "step": 1090
1318
- },
1319
- {
1320
- "epoch": 8.72,
1321
- "step": 1090,
1322
- "train/loss": 3.1708948426967254e-06
1323
- },
1324
- {
1325
- "epoch": 8.8,
1326
- "grad_norm": 0.002471923828125,
1327
- "learning_rate": 5.600000000000001e-06,
1328
- "loss": 0.0003,
1329
- "step": 1100
1330
- },
1331
- {
1332
- "epoch": 8.8,
1333
- "step": 1100,
1334
- "train/loss": 3.5236623716627946e-06
1335
- },
1336
- {
1337
- "epoch": 8.88,
1338
- "grad_norm": 0.00616455078125,
1339
- "learning_rate": 5.560000000000001e-06,
1340
- "loss": 0.0001,
1341
- "step": 1110
1342
- },
1343
- {
1344
- "epoch": 8.88,
1345
- "step": 1110,
1346
- "train/loss": 8.851275197230279e-07
1347
- },
1348
- {
1349
- "epoch": 8.96,
1350
- "grad_norm": 0.0020599365234375,
1351
- "learning_rate": 5.5200000000000005e-06,
1352
- "loss": 0.0001,
1353
- "step": 1120
1354
- },
1355
- {
1356
- "epoch": 8.96,
1357
- "step": 1120,
1358
- "train/loss": 1.1822639862657525e-05
1359
- },
1360
- {
1361
- "epoch": 9.04,
1362
- "grad_norm": 0.0030059814453125,
1363
- "learning_rate": 5.480000000000001e-06,
1364
- "loss": 0.0002,
1365
- "step": 1130
1366
- },
1367
- {
1368
- "epoch": 9.04,
1369
- "step": 1130,
1370
- "train/loss": 6.532348834298318e-06
1371
- },
1372
- {
1373
- "epoch": 9.12,
1374
- "grad_norm": 0.005523681640625,
1375
- "learning_rate": 5.4400000000000004e-06,
1376
- "loss": 0.0,
1377
- "step": 1140
1378
- },
1379
- {
1380
- "epoch": 9.12,
1381
- "step": 1140,
1382
- "train/loss": 1.3127173588145524e-05
1383
- },
1384
- {
1385
- "epoch": 9.2,
1386
- "grad_norm": 0.00994873046875,
1387
- "learning_rate": 5.400000000000001e-06,
1388
- "loss": 0.0001,
1389
- "step": 1150
1390
- },
1391
- {
1392
- "epoch": 9.2,
1393
- "step": 1150,
1394
- "train/loss": 4.936652203468839e-06
1395
- },
1396
- {
1397
- "epoch": 9.28,
1398
- "grad_norm": 0.00124359130859375,
1399
- "learning_rate": 5.36e-06,
1400
- "loss": 0.0002,
1401
- "step": 1160
1402
- },
1403
- {
1404
- "epoch": 9.28,
1405
- "step": 1160,
1406
- "train/loss": 2.0295181002438767e-06
1407
- },
1408
- {
1409
- "epoch": 9.36,
1410
- "grad_norm": 0.01446533203125,
1411
- "learning_rate": 5.320000000000001e-06,
1412
- "loss": 0.0001,
1413
- "step": 1170
1414
- },
1415
- {
1416
- "epoch": 9.36,
1417
- "step": 1170,
1418
- "train/loss": 5.721064553654287e-06
1419
- },
1420
- {
1421
- "epoch": 9.44,
1422
- "grad_norm": 0.0029296875,
1423
- "learning_rate": 5.28e-06,
1424
- "loss": 0.0,
1425
- "step": 1180
1426
- },
1427
- {
1428
- "epoch": 9.44,
1429
- "step": 1180,
1430
- "train/loss": 0.00014855993504170328
1431
- },
1432
- {
1433
- "epoch": 9.52,
1434
- "grad_norm": 0.0234375,
1435
- "learning_rate": 5.240000000000001e-06,
1436
- "loss": 0.0001,
1437
- "step": 1190
1438
- },
1439
- {
1440
- "epoch": 9.52,
1441
- "step": 1190,
1442
- "train/loss": 1.1649264706647955e-05
1443
- },
1444
- {
1445
- "epoch": 9.6,
1446
- "grad_norm": 0.000957489013671875,
1447
- "learning_rate": 5.2e-06,
1448
- "loss": 0.0001,
1449
- "step": 1200
1450
- },
1451
- {
1452
- "epoch": 9.6,
1453
- "step": 1200,
1454
- "train/loss": 5.042222710471833e-06
1455
- },
1456
- {
1457
- "epoch": 9.68,
1458
- "grad_norm": 0.001708984375,
1459
- "learning_rate": 5.1600000000000006e-06,
1460
- "loss": 0.0,
1461
- "step": 1210
1462
- },
1463
- {
1464
- "epoch": 9.68,
1465
- "step": 1210,
1466
- "train/loss": 1.6043163668655325e-06
1467
- },
1468
- {
1469
- "epoch": 9.76,
1470
- "grad_norm": 0.0021514892578125,
1471
- "learning_rate": 5.12e-06,
1472
- "loss": 0.0,
1473
- "step": 1220
1474
- },
1475
- {
1476
- "epoch": 9.76,
1477
- "step": 1220,
1478
- "train/loss": 1.9052860125157167e-06
1479
- },
1480
- {
1481
- "epoch": 9.84,
1482
- "grad_norm": 0.0303955078125,
1483
- "learning_rate": 5.0800000000000005e-06,
1484
- "loss": 0.0,
1485
- "step": 1230
1486
- },
1487
- {
1488
- "epoch": 9.84,
1489
- "step": 1230,
1490
- "train/loss": 1.1143329174956307e-05
1491
- },
1492
- {
1493
- "epoch": 9.92,
1494
- "grad_norm": 0.062255859375,
1495
- "learning_rate": 5.04e-06,
1496
- "loss": 0.0,
1497
- "step": 1240
1498
- },
1499
- {
1500
- "epoch": 9.92,
1501
- "step": 1240,
1502
- "train/loss": 4.859384262090316e-06
1503
- },
1504
- {
1505
- "epoch": 10.0,
1506
- "grad_norm": 0.0390625,
1507
- "learning_rate": 5e-06,
1508
- "loss": 0.0,
1509
- "step": 1250
1510
- },
1511
- {
1512
- "epoch": 10.0,
1513
- "step": 1250,
1514
- "train/loss": 1.2310964848438744e-05
1515
- },
1516
- {
1517
- "epoch": 10.08,
1518
- "grad_norm": 0.0096435546875,
1519
- "learning_rate": 4.960000000000001e-06,
1520
- "loss": 0.0,
1521
- "step": 1260
1522
- },
1523
- {
1524
- "epoch": 10.08,
1525
- "step": 1260,
1526
- "train/loss": 3.4872434753197012e-06
1527
- },
1528
- {
1529
- "epoch": 10.16,
1530
- "grad_norm": 0.007354736328125,
1531
- "learning_rate": 4.92e-06,
1532
- "loss": 0.0,
1533
- "step": 1270
1534
- },
1535
- {
1536
- "epoch": 10.16,
1537
- "step": 1270,
1538
- "train/loss": 2.7374915589462034e-05
1539
- },
1540
- {
1541
- "epoch": 10.24,
1542
- "grad_norm": 0.0050048828125,
1543
- "learning_rate": 4.880000000000001e-06,
1544
- "loss": 0.0,
1545
- "step": 1280
1546
- },
1547
- {
1548
- "epoch": 10.24,
1549
- "step": 1280,
1550
- "train/loss": 9.079674896383949e-07
1551
- },
1552
- {
1553
- "epoch": 10.32,
1554
- "grad_norm": 0.0167236328125,
1555
- "learning_rate": 4.84e-06,
1556
- "loss": 0.0,
1557
- "step": 1290
1558
- },
1559
- {
1560
- "epoch": 10.32,
1561
- "step": 1290,
1562
- "train/loss": 6.887464405735955e-05
1563
- },
1564
- {
1565
- "epoch": 10.4,
1566
- "grad_norm": 0.0030059814453125,
1567
- "learning_rate": 4.800000000000001e-06,
1568
- "loss": 0.0,
1569
- "step": 1300
1570
- },
1571
- {
1572
- "epoch": 10.4,
1573
- "step": 1300,
1574
- "train/loss": 4.588063347910065e-06
1575
- },
1576
- {
1577
- "epoch": 10.48,
1578
- "grad_norm": 0.0810546875,
1579
- "learning_rate": 4.76e-06,
1580
- "loss": 0.0,
1581
- "step": 1310
1582
- },
1583
- {
1584
- "epoch": 10.48,
1585
- "step": 1310,
1586
- "train/loss": 3.869086867780425e-06
1587
- },
1588
- {
1589
- "epoch": 10.56,
1590
- "grad_norm": 0.00494384765625,
1591
- "learning_rate": 4.7200000000000005e-06,
1592
- "loss": 0.0002,
1593
- "step": 1320
1594
- },
1595
- {
1596
- "epoch": 10.56,
1597
- "step": 1320,
1598
- "train/loss": 7.87773046795337e-07
1599
- },
1600
- {
1601
- "epoch": 10.64,
1602
- "grad_norm": 0.01495361328125,
1603
- "learning_rate": 4.680000000000001e-06,
1604
- "loss": 0.0,
1605
- "step": 1330
1606
- },
1607
- {
1608
- "epoch": 10.64,
1609
- "step": 1330,
1610
- "train/loss": 2.0062170733581297e-05
1611
- },
1612
- {
1613
- "epoch": 10.72,
1614
- "grad_norm": 0.00360107421875,
1615
- "learning_rate": 4.6400000000000005e-06,
1616
- "loss": 0.0,
1617
- "step": 1340
1618
- },
1619
- {
1620
- "epoch": 10.72,
1621
- "step": 1340,
1622
- "train/loss": 3.143127059956896e-06
1623
- },
1624
- {
1625
- "epoch": 10.8,
1626
- "grad_norm": 0.003448486328125,
1627
- "learning_rate": 4.600000000000001e-06,
1628
- "loss": 0.0,
1629
- "step": 1350
1630
- },
1631
- {
1632
- "epoch": 10.8,
1633
- "step": 1350,
1634
- "train/loss": 6.601278528250987e-06
1635
- },
1636
- {
1637
- "epoch": 10.88,
1638
- "grad_norm": 0.003021240234375,
1639
- "learning_rate": 4.56e-06,
1640
- "loss": 0.0003,
1641
- "step": 1360
1642
- },
1643
- {
1644
- "epoch": 10.88,
1645
- "step": 1360,
1646
- "train/loss": 4.25061580244801e-06
1647
- },
1648
- {
1649
- "epoch": 10.96,
1650
- "grad_norm": 0.00433349609375,
1651
- "learning_rate": 4.520000000000001e-06,
1652
- "loss": 0.0001,
1653
- "step": 1370
1654
- },
1655
- {
1656
- "epoch": 10.96,
1657
- "step": 1370,
1658
- "train/loss": 2.1255189039948164e-06
1659
- },
1660
- {
1661
- "epoch": 11.04,
1662
- "grad_norm": 0.06005859375,
1663
- "learning_rate": 4.48e-06,
1664
- "loss": 0.0,
1665
- "step": 1380
1666
- },
1667
- {
1668
- "epoch": 11.04,
1669
- "step": 1380,
1670
- "train/loss": 2.866860768335755e-06
1671
- },
1672
- {
1673
- "epoch": 11.12,
1674
- "grad_norm": 0.0167236328125,
1675
- "learning_rate": 4.440000000000001e-06,
1676
- "loss": 0.0,
1677
- "step": 1390
1678
- },
1679
- {
1680
- "epoch": 11.12,
1681
- "step": 1390,
1682
- "train/loss": 1.4920918829375296e-06
1683
- },
1684
- {
1685
- "epoch": 11.2,
1686
- "grad_norm": 0.185546875,
1687
- "learning_rate": 4.4e-06,
1688
- "loss": 0.0004,
1689
- "step": 1400
1690
- },
1691
- {
1692
- "epoch": 11.2,
1693
- "step": 1400,
1694
- "train/loss": 1.3183645251046983e-06
1695
- },
1696
- {
1697
- "epoch": 11.28,
1698
- "grad_norm": 0.0244140625,
1699
- "learning_rate": 4.360000000000001e-06,
1700
- "loss": 0.0,
1701
- "step": 1410
1702
- },
1703
- {
1704
- "epoch": 11.28,
1705
- "step": 1410,
1706
- "train/loss": 2.6989214347850066e-06
1707
- },
1708
- {
1709
- "epoch": 11.36,
1710
- "grad_norm": 0.0087890625,
1711
- "learning_rate": 4.32e-06,
1712
- "loss": 0.0004,
1713
- "step": 1420
1714
- },
1715
- {
1716
- "epoch": 11.36,
1717
- "step": 1420,
1718
- "train/loss": 2.1954119802103378e-06
1719
- },
1720
- {
1721
- "epoch": 11.44,
1722
- "grad_norm": 0.109375,
1723
- "learning_rate": 4.2800000000000005e-06,
1724
- "loss": 0.0,
1725
- "step": 1430
1726
- },
1727
- {
1728
- "epoch": 11.44,
1729
- "step": 1430,
1730
- "train/loss": 1.8517592934585991e-06
1731
- },
1732
- {
1733
- "epoch": 11.52,
1734
- "grad_norm": 0.0027923583984375,
1735
- "learning_rate": 4.24e-06,
1736
- "loss": 0.0,
1737
- "step": 1440
1738
- },
1739
- {
1740
- "epoch": 11.52,
1741
- "step": 1440,
1742
- "train/loss": 6.585671144421212e-06
1743
- },
1744
- {
1745
- "epoch": 11.6,
1746
- "grad_norm": 0.0263671875,
1747
- "learning_rate": 4.2000000000000004e-06,
1748
- "loss": 0.0,
1749
- "step": 1450
1750
- },
1751
- {
1752
- "epoch": 11.6,
1753
- "step": 1450,
1754
- "train/loss": 2.106990450556623e-06
1755
- },
1756
- {
1757
- "epoch": 11.68,
1758
- "grad_norm": 0.00159454345703125,
1759
- "learning_rate": 4.16e-06,
1760
- "loss": 0.0,
1761
- "step": 1460
1762
- },
1763
- {
1764
- "epoch": 11.68,
1765
- "step": 1460,
1766
- "train/loss": 1.052016614266904e-06
1767
- },
1768
- {
1769
- "epoch": 11.76,
1770
- "grad_norm": 0.375,
1771
- "learning_rate": 4.12e-06,
1772
- "loss": 0.0001,
1773
- "step": 1470
1774
- },
1775
- {
1776
- "epoch": 11.76,
1777
- "step": 1470,
1778
- "train/loss": 1.0533399290579837e-05
1779
- },
1780
- {
1781
- "epoch": 11.84,
1782
- "grad_norm": 0.01312255859375,
1783
- "learning_rate": 4.08e-06,
1784
- "loss": 0.0001,
1785
- "step": 1480
1786
- },
1787
- {
1788
- "epoch": 11.84,
1789
- "step": 1480,
1790
- "train/loss": 6.241708433663007e-06
1791
- },
1792
- {
1793
- "epoch": 11.92,
1794
- "grad_norm": 0.00494384765625,
1795
- "learning_rate": 4.04e-06,
1796
- "loss": 0.0,
1797
- "step": 1490
1798
- },
1799
- {
1800
- "epoch": 11.92,
1801
- "step": 1490,
1802
- "train/loss": 4.304019512346713e-06
1803
- },
1804
- {
1805
- "epoch": 12.0,
1806
- "grad_norm": 0.004364013671875,
1807
- "learning_rate": 4.000000000000001e-06,
1808
- "loss": 0.0001,
1809
- "step": 1500
1810
- },
1811
- {
1812
- "epoch": 12.0,
1813
- "step": 1500,
1814
- "train/loss": 6.725377375005337e-07
1815
- },
1816
- {
1817
- "epoch": 12.08,
1818
- "grad_norm": 0.0279541015625,
1819
- "learning_rate": 3.96e-06,
1820
- "loss": 0.0001,
1821
- "step": 1510
1822
- },
1823
- {
1824
- "epoch": 12.08,
1825
- "step": 1510,
1826
- "train/loss": 1.8328212263440946e-06
1827
- },
1828
- {
1829
- "epoch": 12.16,
1830
- "grad_norm": 0.003021240234375,
1831
- "learning_rate": 3.920000000000001e-06,
1832
- "loss": 0.0,
1833
- "step": 1520
1834
- },
1835
- {
1836
- "epoch": 12.16,
1837
- "step": 1520,
1838
- "train/loss": 1.849480759119615e-05
1839
- },
1840
- {
1841
- "epoch": 12.24,
1842
- "grad_norm": 0.0673828125,
1843
- "learning_rate": 3.88e-06,
1844
- "loss": 0.0,
1845
- "step": 1530
1846
- },
1847
- {
1848
- "epoch": 12.24,
1849
- "step": 1530,
1850
- "train/loss": 9.769206371856853e-06
1851
- },
1852
- {
1853
- "epoch": 12.32,
1854
- "grad_norm": 0.060546875,
1855
- "learning_rate": 3.8400000000000005e-06,
1856
- "loss": 0.0,
1857
- "step": 1540
1858
- },
1859
- {
1860
- "epoch": 12.32,
1861
- "step": 1540,
1862
- "train/loss": 4.941887254972244e-06
1863
- },
1864
- {
1865
- "epoch": 12.4,
1866
- "grad_norm": 0.002044677734375,
1867
- "learning_rate": 3.8000000000000005e-06,
1868
- "loss": 0.0,
1869
- "step": 1550
1870
- },
1871
- {
1872
- "epoch": 12.4,
1873
- "step": 1550,
1874
- "train/loss": 0.0008674569544382393
1875
- },
1876
- {
1877
- "epoch": 12.48,
1878
- "grad_norm": 0.00787353515625,
1879
- "learning_rate": 3.7600000000000004e-06,
1880
- "loss": 0.0,
1881
- "step": 1560
1882
- },
1883
- {
1884
- "epoch": 12.48,
1885
- "step": 1560,
1886
- "train/loss": 5.4637007451674435e-06
1887
- },
1888
- {
1889
- "epoch": 12.56,
1890
- "grad_norm": 0.0030059814453125,
1891
- "learning_rate": 3.7200000000000004e-06,
1892
- "loss": 0.0,
1893
- "step": 1570
1894
- },
1895
- {
1896
- "epoch": 12.56,
1897
- "step": 1570,
1898
- "train/loss": 1.8032819752988871e-06
1899
- },
1900
- {
1901
- "epoch": 12.64,
1902
- "grad_norm": 0.000843048095703125,
1903
- "learning_rate": 3.6800000000000003e-06,
1904
- "loss": 0.0001,
1905
- "step": 1580
1906
- },
1907
- {
1908
- "epoch": 12.64,
1909
- "step": 1580,
1910
- "train/loss": 1.6003571090550395e-06
1911
- },
1912
- {
1913
- "epoch": 12.72,
1914
- "grad_norm": 0.00531005859375,
1915
- "learning_rate": 3.6400000000000003e-06,
1916
- "loss": 0.0,
1917
- "step": 1590
1918
- },
1919
- {
1920
- "epoch": 12.72,
1921
- "step": 1590,
1922
- "train/loss": 7.059289600874763e-06
1923
- },
1924
- {
1925
- "epoch": 12.8,
1926
- "grad_norm": 0.0018157958984375,
1927
- "learning_rate": 3.6000000000000003e-06,
1928
- "loss": 0.0,
1929
- "step": 1600
1930
- },
1931
- {
1932
- "epoch": 12.8,
1933
- "step": 1600,
1934
- "train/loss": 1.4950710465200245e-06
1935
- },
1936
- {
1937
- "epoch": 12.88,
1938
- "grad_norm": 0.00408935546875,
1939
- "learning_rate": 3.5600000000000002e-06,
1940
- "loss": 0.0,
1941
- "step": 1610
1942
- },
1943
- {
1944
- "epoch": 12.88,
1945
- "step": 1610,
1946
- "train/loss": 4.470039584703045e-06
1947
- },
1948
- {
1949
- "epoch": 12.96,
1950
- "grad_norm": 0.006195068359375,
1951
- "learning_rate": 3.52e-06,
1952
- "loss": 0.0001,
1953
- "step": 1620
1954
- },
1955
- {
1956
- "epoch": 12.96,
1957
- "step": 1620,
1958
- "train/loss": 1.7133684195869137e-06
1959
- },
1960
- {
1961
- "epoch": 13.04,
1962
- "grad_norm": 0.2041015625,
1963
- "learning_rate": 3.48e-06,
1964
- "loss": 0.0001,
1965
- "step": 1630
1966
- },
1967
- {
1968
- "epoch": 13.04,
1969
- "step": 1630,
1970
- "train/loss": 1.756774690875318e-05
1971
- },
1972
- {
1973
- "epoch": 13.12,
1974
- "grad_norm": 0.01397705078125,
1975
- "learning_rate": 3.44e-06,
1976
- "loss": 0.0,
1977
- "step": 1640
1978
- },
1979
- {
1980
- "epoch": 13.12,
1981
- "step": 1640,
1982
- "train/loss": 1.2655867749344907e-06
1983
- },
1984
- {
1985
- "epoch": 13.2,
1986
- "grad_norm": 0.020751953125,
1987
- "learning_rate": 3.4000000000000005e-06,
1988
- "loss": 0.0001,
1989
- "step": 1650
1990
- },
1991
- {
1992
- "epoch": 13.2,
1993
- "step": 1650,
1994
- "train/loss": 4.243750026944326e-06
1995
- },
1996
- {
1997
- "epoch": 13.28,
1998
- "grad_norm": 0.00469970703125,
1999
- "learning_rate": 3.3600000000000004e-06,
2000
- "loss": 0.0,
2001
- "step": 1660
2002
- },
2003
- {
2004
- "epoch": 13.28,
2005
- "step": 1660,
2006
- "train/loss": 8.94761978997849e-06
2007
- },
2008
- {
2009
- "epoch": 13.36,
2010
- "grad_norm": 0.0027923583984375,
2011
- "learning_rate": 3.3200000000000004e-06,
2012
- "loss": 0.0,
2013
- "step": 1670
2014
- },
2015
- {
2016
- "epoch": 13.36,
2017
- "step": 1670,
2018
- "train/loss": 1.1165860769324354e-06
2019
- },
2020
- {
2021
- "epoch": 13.44,
2022
- "grad_norm": 0.0281982421875,
2023
- "learning_rate": 3.2800000000000004e-06,
2024
- "loss": 0.0,
2025
- "step": 1680
2026
- },
2027
- {
2028
- "epoch": 13.44,
2029
- "step": 1680,
2030
- "train/loss": 2.4268495053547667e-06
2031
- },
2032
- {
2033
- "epoch": 13.52,
2034
- "grad_norm": 0.00140380859375,
2035
- "learning_rate": 3.2400000000000003e-06,
2036
- "loss": 0.0,
2037
- "step": 1690
2038
- },
2039
- {
2040
- "epoch": 13.52,
2041
- "step": 1690,
2042
- "train/loss": 8.542659634258598e-05
2043
- },
2044
- {
2045
- "epoch": 13.6,
2046
- "grad_norm": 0.0022430419921875,
2047
- "learning_rate": 3.2000000000000003e-06,
2048
- "loss": 0.0,
2049
- "step": 1700
2050
- },
2051
- {
2052
- "epoch": 13.6,
2053
- "step": 1700,
2054
- "train/loss": 8.702247100700333e-07
2055
- },
2056
- {
2057
- "epoch": 13.68,
2058
- "grad_norm": 0.01507568359375,
2059
- "learning_rate": 3.1600000000000002e-06,
2060
- "loss": 0.0,
2061
- "step": 1710
2062
- },
2063
- {
2064
- "epoch": 13.68,
2065
- "step": 1710,
2066
- "train/loss": 1.119564558393904e-06
2067
- },
2068
- {
2069
- "epoch": 13.76,
2070
- "grad_norm": 0.2470703125,
2071
- "learning_rate": 3.12e-06,
2072
- "loss": 0.0001,
2073
- "step": 1720
2074
- },
2075
- {
2076
- "epoch": 13.76,
2077
- "step": 1720,
2078
- "train/loss": 4.818349225388374e-06
2079
- },
2080
- {
2081
- "epoch": 13.84,
2082
- "grad_norm": 0.10791015625,
2083
- "learning_rate": 3.08e-06,
2084
- "loss": 0.0,
2085
- "step": 1730
2086
- },
2087
- {
2088
- "epoch": 13.84,
2089
- "step": 1730,
2090
- "train/loss": 2.6204995720036095e-06
2091
- },
2092
- {
2093
- "epoch": 13.92,
2094
- "grad_norm": 0.002899169921875,
2095
- "learning_rate": 3.04e-06,
2096
- "loss": 0.0,
2097
- "step": 1740
2098
- },
2099
- {
2100
- "epoch": 13.92,
2101
- "step": 1740,
2102
- "train/loss": 1.994286685658153e-05
2103
- },
2104
- {
2105
- "epoch": 14.0,
2106
- "grad_norm": 0.064453125,
2107
- "learning_rate": 3e-06,
2108
- "loss": 0.0,
2109
- "step": 1750
2110
- },
2111
- {
2112
- "epoch": 14.0,
2113
- "step": 1750,
2114
- "train/loss": 6.559032954100985e-06
2115
- },
2116
- {
2117
- "epoch": 14.08,
2118
- "grad_norm": 0.00469970703125,
2119
- "learning_rate": 2.96e-06,
2120
- "loss": 0.0,
2121
- "step": 1760
2122
- },
2123
- {
2124
- "epoch": 14.08,
2125
- "step": 1760,
2126
- "train/loss": 1.7702242303130333e-06
2127
- },
2128
- {
2129
- "epoch": 14.16,
2130
- "grad_norm": 0.01202392578125,
2131
- "learning_rate": 2.92e-06,
2132
- "loss": 0.0,
2133
- "step": 1770
2134
- },
2135
- {
2136
- "epoch": 14.16,
2137
- "step": 1770,
2138
- "train/loss": 3.1200149805954425e-06
2139
- },
2140
- {
2141
- "epoch": 14.24,
2142
- "grad_norm": 0.0019683837890625,
2143
- "learning_rate": 2.88e-06,
2144
- "loss": 0.0,
2145
- "step": 1780
2146
- },
2147
- {
2148
- "epoch": 14.24,
2149
- "step": 1780,
2150
- "train/loss": 1.8795326468534768e-05
2151
- },
2152
- {
2153
- "epoch": 14.32,
2154
- "grad_norm": 0.00323486328125,
2155
- "learning_rate": 2.84e-06,
2156
- "loss": 0.0,
2157
- "step": 1790
2158
- },
2159
- {
2160
- "epoch": 14.32,
2161
- "step": 1790,
2162
- "train/loss": 1.5069895198394079e-06
2163
- },
2164
- {
2165
- "epoch": 14.4,
2166
- "grad_norm": 0.0537109375,
2167
- "learning_rate": 2.8000000000000003e-06,
2168
- "loss": 0.0,
2169
- "step": 1800
2170
- },
2171
- {
2172
- "epoch": 14.4,
2173
- "step": 1800,
2174
- "train/loss": 2.0881050204479834e-06
2175
- },
2176
- {
2177
- "epoch": 14.48,
2178
- "grad_norm": 0.1982421875,
2179
- "learning_rate": 2.7600000000000003e-06,
2180
- "loss": 0.0001,
2181
- "step": 1810
2182
- },
2183
- {
2184
- "epoch": 14.48,
2185
- "step": 1810,
2186
- "train/loss": 6.848983503005002e-06
2187
- },
2188
- {
2189
- "epoch": 14.56,
2190
- "grad_norm": 0.0233154296875,
2191
- "learning_rate": 2.7200000000000002e-06,
2192
- "loss": 0.0001,
2193
- "step": 1820
2194
- },
2195
- {
2196
- "epoch": 14.56,
2197
- "step": 1820,
2198
- "train/loss": 3.0963235531089595e-06
2199
- },
2200
- {
2201
- "epoch": 14.64,
2202
- "grad_norm": 0.000762939453125,
2203
- "learning_rate": 2.68e-06,
2204
- "loss": 0.0,
2205
- "step": 1830
2206
- },
2207
- {
2208
- "epoch": 14.64,
2209
- "step": 1830,
2210
- "train/loss": 1.3718830587094999e-06
2211
- },
2212
- {
2213
- "epoch": 14.72,
2214
- "grad_norm": 0.0057373046875,
2215
- "learning_rate": 2.64e-06,
2216
- "loss": 0.0,
2217
- "step": 1840
2218
- },
2219
- {
2220
- "epoch": 14.72,
2221
- "step": 1840,
2222
- "train/loss": 1.155993049906101e-05
2223
- },
2224
- {
2225
- "epoch": 14.8,
2226
- "grad_norm": 0.05810546875,
2227
- "learning_rate": 2.6e-06,
2228
- "loss": 0.0001,
2229
- "step": 1850
2230
- },
2231
- {
2232
- "epoch": 14.8,
2233
- "step": 1850,
2234
- "train/loss": 0.0021582189947366714
2235
- },
2236
- {
2237
- "epoch": 14.88,
2238
- "grad_norm": 0.01068115234375,
2239
- "learning_rate": 2.56e-06,
2240
- "loss": 0.0001,
2241
- "step": 1860
2242
- },
2243
- {
2244
- "epoch": 14.88,
2245
- "step": 1860,
2246
- "train/loss": 8.173182322934736e-06
2247
- },
2248
- {
2249
- "epoch": 14.96,
2250
- "grad_norm": 0.000827789306640625,
2251
- "learning_rate": 2.52e-06,
2252
- "loss": 0.0,
2253
- "step": 1870
2254
- },
2255
- {
2256
- "epoch": 14.96,
2257
- "step": 1870,
2258
- "train/loss": 1.2834279004891869e-05
2259
- },
2260
- {
2261
- "epoch": 15.04,
2262
- "grad_norm": 0.06591796875,
2263
- "learning_rate": 2.4800000000000004e-06,
2264
- "loss": 0.0,
2265
- "step": 1880
2266
- },
2267
- {
2268
- "epoch": 15.04,
2269
- "step": 1880,
2270
- "train/loss": 0.0005998230190016329
2271
- },
2272
- {
2273
- "epoch": 15.12,
2274
- "grad_norm": 0.004669189453125,
2275
- "learning_rate": 2.4400000000000004e-06,
2276
- "loss": 0.0,
2277
- "step": 1890
2278
- },
2279
- {
2280
- "epoch": 15.12,
2281
- "step": 1890,
2282
- "train/loss": 1.5745865312055685e-05
2283
- },
2284
- {
2285
- "epoch": 15.2,
2286
- "grad_norm": 0.02294921875,
2287
- "learning_rate": 2.4000000000000003e-06,
2288
- "loss": 0.0,
2289
- "step": 1900
2290
- },
2291
- {
2292
- "epoch": 15.2,
2293
- "step": 1900,
2294
- "train/loss": 2.679145381989656e-06
2295
- },
2296
- {
2297
- "epoch": 15.28,
2298
- "grad_norm": 0.007659912109375,
2299
- "learning_rate": 2.3600000000000003e-06,
2300
- "loss": 0.0,
2301
- "step": 1910
2302
- },
2303
- {
2304
- "epoch": 15.28,
2305
- "step": 1910,
2306
- "train/loss": 7.805366476532072e-05
2307
- },
2308
- {
2309
- "epoch": 15.36,
2310
- "grad_norm": 0.00213623046875,
2311
- "learning_rate": 2.3200000000000002e-06,
2312
- "loss": 0.0001,
2313
- "step": 1920
2314
- },
2315
- {
2316
- "epoch": 15.36,
2317
- "step": 1920,
2318
- "train/loss": 3.0448240977420937e-06
2319
- },
2320
- {
2321
- "epoch": 15.44,
2322
- "grad_norm": 0.04541015625,
2323
- "learning_rate": 2.28e-06,
2324
- "loss": 0.0,
2325
- "step": 1930
2326
- },
2327
- {
2328
- "epoch": 15.44,
2329
- "step": 1930,
2330
- "train/loss": 8.960526542978187e-07
2331
- },
2332
- {
2333
- "epoch": 15.52,
2334
- "grad_norm": 0.0021820068359375,
2335
- "learning_rate": 2.24e-06,
2336
- "loss": 0.0,
2337
- "step": 1940
2338
- },
2339
- {
2340
- "epoch": 15.52,
2341
- "step": 1940,
2342
- "train/loss": 6.716804364259588e-06
2343
- },
2344
- {
2345
- "epoch": 15.6,
2346
- "grad_norm": 0.0021209716796875,
2347
- "learning_rate": 2.2e-06,
2348
- "loss": 0.0,
2349
- "step": 1950
2350
- },
2351
- {
2352
- "epoch": 15.6,
2353
- "step": 1950,
2354
- "train/loss": 7.708176417509094e-05
2355
- },
2356
- {
2357
- "epoch": 15.68,
2358
- "grad_norm": 0.006103515625,
2359
- "learning_rate": 2.16e-06,
2360
- "loss": 0.0,
2361
- "step": 1960
2362
- },
2363
- {
2364
- "epoch": 15.68,
2365
- "step": 1960,
2366
- "train/loss": 2.2430820081353886e-06
2367
- },
2368
- {
2369
- "epoch": 15.76,
2370
- "grad_norm": 0.0059814453125,
2371
- "learning_rate": 2.12e-06,
2372
- "loss": 0.0,
2373
- "step": 1970
2374
- },
2375
- {
2376
- "epoch": 15.76,
2377
- "step": 1970,
2378
- "train/loss": 1.1175829968124162e-06
2379
- },
2380
- {
2381
- "epoch": 15.84,
2382
- "grad_norm": 0.0263671875,
2383
- "learning_rate": 2.08e-06,
2384
- "loss": 0.0,
2385
- "step": 1980
2386
- },
2387
- {
2388
- "epoch": 15.84,
2389
- "step": 1980,
2390
- "train/loss": 3.1032529932417674e-06
2391
- },
2392
- {
2393
- "epoch": 15.92,
2394
- "grad_norm": 0.004425048828125,
2395
- "learning_rate": 2.04e-06,
2396
- "loss": 0.0001,
2397
- "step": 1990
2398
- },
2399
- {
2400
- "epoch": 15.92,
2401
- "step": 1990,
2402
- "train/loss": 1.1532710232131649e-05
2403
- },
2404
- {
2405
- "epoch": 16.0,
2406
- "grad_norm": 0.04150390625,
2407
- "learning_rate": 2.0000000000000003e-06,
2408
- "loss": 0.0001,
2409
- "step": 2000
2410
- },
2411
- {
2412
- "epoch": 16.0,
2413
- "step": 2000,
2414
- "train/loss": 1.6921478618314723e-06
2415
- },
2416
- {
2417
- "epoch": 16.08,
2418
- "grad_norm": 0.006500244140625,
2419
- "learning_rate": 1.9600000000000003e-06,
2420
- "loss": 0.0,
2421
- "step": 2010
2422
- },
2423
- {
2424
- "epoch": 16.08,
2425
- "step": 2010,
2426
- "train/loss": 1.5082887330208905e-06
2427
- },
2428
- {
2429
- "epoch": 16.16,
2430
- "grad_norm": 0.024658203125,
2431
- "learning_rate": 1.9200000000000003e-06,
2432
- "loss": 0.0,
2433
- "step": 2020
2434
- },
2435
- {
2436
- "epoch": 16.16,
2437
- "step": 2020,
2438
- "train/loss": 2.4297712570842123e-06
2439
- },
2440
- {
2441
- "epoch": 16.24,
2442
- "grad_norm": 0.0026092529296875,
2443
- "learning_rate": 1.8800000000000002e-06,
2444
- "loss": 0.0,
2445
- "step": 2030
2446
- },
2447
- {
2448
- "epoch": 16.24,
2449
- "step": 2030,
2450
- "train/loss": 0.00027944811154156923
2451
- },
2452
- {
2453
- "epoch": 16.32,
2454
- "grad_norm": 0.01434326171875,
2455
- "learning_rate": 1.8400000000000002e-06,
2456
- "loss": 0.0,
2457
- "step": 2040
2458
- },
2459
- {
2460
- "epoch": 16.32,
2461
- "step": 2040,
2462
- "train/loss": 3.3982730656134663e-06
2463
- },
2464
- {
2465
- "epoch": 16.4,
2466
- "grad_norm": 0.00543212890625,
2467
- "learning_rate": 1.8000000000000001e-06,
2468
- "loss": 0.0,
2469
- "step": 2050
2470
- },
2471
- {
2472
- "epoch": 16.4,
2473
- "step": 2050,
2474
- "train/loss": 2.168583250750089e-06
2475
- },
2476
- {
2477
- "epoch": 16.48,
2478
- "grad_norm": 0.044189453125,
2479
- "learning_rate": 1.76e-06,
2480
- "loss": 0.0,
2481
- "step": 2060
2482
- },
2483
- {
2484
- "epoch": 16.48,
2485
- "step": 2060,
2486
- "train/loss": 1.3261948197396123e-06
2487
- },
2488
- {
2489
- "epoch": 16.56,
2490
- "grad_norm": 0.01324462890625,
2491
- "learning_rate": 1.72e-06,
2492
- "loss": 0.0,
2493
- "step": 2070
2494
- },
2495
- {
2496
- "epoch": 16.56,
2497
- "step": 2070,
2498
- "train/loss": 0.00011921257100766525
2499
- },
2500
- {
2501
- "epoch": 16.64,
2502
- "grad_norm": 0.00994873046875,
2503
- "learning_rate": 1.6800000000000002e-06,
2504
- "loss": 0.0001,
2505
- "step": 2080
2506
- },
2507
- {
2508
- "epoch": 16.64,
2509
- "step": 2080,
2510
- "train/loss": 5.616849648504285e-06
2511
- },
2512
- {
2513
- "epoch": 16.72,
2514
- "grad_norm": 0.0019378662109375,
2515
- "learning_rate": 1.6400000000000002e-06,
2516
- "loss": 0.0,
2517
- "step": 2090
2518
- },
2519
- {
2520
- "epoch": 16.72,
2521
- "step": 2090,
2522
- "train/loss": 2.1993721475155326e-06
2523
- },
2524
- {
2525
- "epoch": 16.8,
2526
- "grad_norm": 0.205078125,
2527
- "learning_rate": 1.6000000000000001e-06,
2528
- "loss": 0.0001,
2529
- "step": 2100
2530
- },
2531
- {
2532
- "epoch": 16.8,
2533
- "step": 2100,
2534
- "train/loss": 4.359588729130337e-06
2535
- },
2536
- {
2537
- "epoch": 16.88,
2538
- "grad_norm": 0.0081787109375,
2539
- "learning_rate": 1.56e-06,
2540
- "loss": 0.0,
2541
- "step": 2110
2542
- },
2543
- {
2544
- "epoch": 16.88,
2545
- "step": 2110,
2546
- "train/loss": 5.8437781262909994e-05
2547
- },
2548
- {
2549
- "epoch": 16.96,
2550
- "grad_norm": 0.005950927734375,
2551
- "learning_rate": 1.52e-06,
2552
- "loss": 0.0,
2553
- "step": 2120
2554
- },
2555
- {
2556
- "epoch": 16.96,
2557
- "step": 2120,
2558
- "train/loss": 2.4972925984911853e-06
2559
- },
2560
- {
2561
- "epoch": 17.04,
2562
- "grad_norm": 0.001068115234375,
2563
- "learning_rate": 1.48e-06,
2564
- "loss": 0.0001,
2565
- "step": 2130
2566
- },
2567
- {
2568
- "epoch": 17.04,
2569
- "step": 2130,
2570
- "train/loss": 2.1669484340236522e-06
2571
- },
2572
- {
2573
- "epoch": 17.12,
2574
- "grad_norm": 0.00677490234375,
2575
- "learning_rate": 1.44e-06,
2576
- "loss": 0.0001,
2577
- "step": 2140
2578
- },
2579
- {
2580
- "epoch": 17.12,
2581
- "step": 2140,
2582
- "train/loss": 1.6788450238891528e-06
2583
- },
2584
- {
2585
- "epoch": 17.2,
2586
- "grad_norm": 0.0030059814453125,
2587
- "learning_rate": 1.4000000000000001e-06,
2588
- "loss": 0.0,
2589
- "step": 2150
2590
- },
2591
- {
2592
- "epoch": 17.2,
2593
- "step": 2150,
2594
- "train/loss": 1.1647519386315253e-05
2595
- },
2596
- {
2597
- "epoch": 17.28,
2598
- "grad_norm": 0.00194549560546875,
2599
- "learning_rate": 1.3600000000000001e-06,
2600
- "loss": 0.0,
2601
- "step": 2160
2602
- },
2603
- {
2604
- "epoch": 17.28,
2605
- "step": 2160,
2606
- "train/loss": 3.008859948749887e-06
2607
- },
2608
- {
2609
- "epoch": 17.36,
2610
- "grad_norm": 0.034423828125,
2611
- "learning_rate": 1.32e-06,
2612
- "loss": 0.0,
2613
- "step": 2170
2614
- },
2615
- {
2616
- "epoch": 17.36,
2617
- "step": 2170,
2618
- "train/loss": 4.579791948344791e-06
2619
- },
2620
- {
2621
- "epoch": 17.44,
2622
- "grad_norm": 0.013916015625,
2623
- "learning_rate": 1.28e-06,
2624
- "loss": 0.0,
2625
- "step": 2180
2626
- },
2627
- {
2628
- "epoch": 17.44,
2629
- "step": 2180,
2630
- "train/loss": 4.607034952641698e-06
2631
- },
2632
- {
2633
- "epoch": 17.52,
2634
- "grad_norm": 0.005859375,
2635
- "learning_rate": 1.2400000000000002e-06,
2636
- "loss": 0.0,
2637
- "step": 2190
2638
- },
2639
- {
2640
- "epoch": 17.52,
2641
- "step": 2190,
2642
- "train/loss": 1.0560316695773508e-05
2643
- },
2644
- {
2645
- "epoch": 17.6,
2646
- "grad_norm": 0.01300048828125,
2647
- "learning_rate": 1.2000000000000002e-06,
2648
- "loss": 0.0,
2649
- "step": 2200
2650
- },
2651
- {
2652
- "epoch": 17.6,
2653
- "step": 2200,
2654
- "train/loss": 5.534459887712728e-06
2655
- },
2656
- {
2657
- "epoch": 17.68,
2658
- "grad_norm": 0.0010986328125,
2659
- "learning_rate": 1.1600000000000001e-06,
2660
- "loss": 0.0,
2661
- "step": 2210
2662
- },
2663
- {
2664
- "epoch": 17.68,
2665
- "step": 2210,
2666
- "train/loss": 7.40236055207788e-06
2667
- },
2668
- {
2669
- "epoch": 17.76,
2670
- "grad_norm": 0.01483154296875,
2671
- "learning_rate": 1.12e-06,
2672
- "loss": 0.0,
2673
- "step": 2220
2674
- },
2675
- {
2676
- "epoch": 17.76,
2677
- "step": 2220,
2678
- "train/loss": 1.2024161151202861e-05
2679
- },
2680
- {
2681
- "epoch": 17.84,
2682
- "grad_norm": 0.044677734375,
2683
- "learning_rate": 1.08e-06,
2684
- "loss": 0.0,
2685
- "step": 2230
2686
- },
2687
- {
2688
- "epoch": 17.84,
2689
- "step": 2230,
2690
- "train/loss": 1.3470184057950974e-05
2691
- },
2692
- {
2693
- "epoch": 17.92,
2694
- "grad_norm": 0.0037689208984375,
2695
- "learning_rate": 1.04e-06,
2696
- "loss": 0.0001,
2697
- "step": 2240
2698
- },
2699
- {
2700
- "epoch": 17.92,
2701
- "step": 2240,
2702
- "train/loss": 1.004335217658081e-06
2703
- },
2704
- {
2705
- "epoch": 18.0,
2706
- "grad_norm": 0.005584716796875,
2707
- "learning_rate": 1.0000000000000002e-06,
2708
- "loss": 0.0,
2709
- "step": 2250
2710
- },
2711
- {
2712
- "epoch": 18.0,
2713
- "step": 2250,
2714
- "train/loss": 9.866492291621398e-06
2715
- },
2716
- {
2717
- "epoch": 18.08,
2718
- "grad_norm": 0.0014495849609375,
2719
- "learning_rate": 9.600000000000001e-07,
2720
- "loss": 0.0001,
2721
- "step": 2260
2722
- },
2723
- {
2724
- "epoch": 18.08,
2725
- "step": 2260,
2726
- "train/loss": 6.16822944721207e-06
2727
- },
2728
- {
2729
- "epoch": 18.16,
2730
- "grad_norm": 0.010986328125,
2731
- "learning_rate": 9.200000000000001e-07,
2732
- "loss": 0.0001,
2733
- "step": 2270
2734
- },
2735
- {
2736
- "epoch": 18.16,
2737
- "step": 2270,
2738
- "train/loss": 6.2418380366580095e-06
2739
- },
2740
- {
2741
- "epoch": 18.24,
2742
- "grad_norm": 0.006927490234375,
2743
- "learning_rate": 8.8e-07,
2744
- "loss": 0.0,
2745
- "step": 2280
2746
- },
2747
- {
2748
- "epoch": 18.24,
2749
- "step": 2280,
2750
- "train/loss": 0.0006859241984784603
2751
- },
2752
- {
2753
- "epoch": 18.32,
2754
- "grad_norm": 0.0185546875,
2755
- "learning_rate": 8.400000000000001e-07,
2756
- "loss": 0.0001,
2757
- "step": 2290
2758
- },
2759
- {
2760
- "epoch": 18.32,
2761
- "step": 2290,
2762
- "train/loss": 1.0311543974239612e-06
2763
- },
2764
- {
2765
- "epoch": 18.4,
2766
- "grad_norm": 0.0054931640625,
2767
- "learning_rate": 8.000000000000001e-07,
2768
- "loss": 0.0,
2769
- "step": 2300
2770
- },
2771
- {
2772
- "epoch": 18.4,
2773
- "step": 2300,
2774
- "train/loss": 4.017131232103566e-06
2775
- },
2776
- {
2777
- "epoch": 18.48,
2778
- "grad_norm": 0.0042724609375,
2779
- "learning_rate": 7.6e-07,
2780
- "loss": 0.0,
2781
- "step": 2310
2782
- },
2783
- {
2784
- "epoch": 18.48,
2785
- "step": 2310,
2786
- "train/loss": 3.2660318538546562e-06
2787
- },
2788
- {
2789
- "epoch": 18.56,
2790
- "grad_norm": 0.0014495849609375,
2791
- "learning_rate": 7.2e-07,
2792
- "loss": 0.0,
2793
- "step": 2320
2794
- },
2795
- {
2796
- "epoch": 18.56,
2797
- "step": 2320,
2798
- "train/loss": 3.985218427260406e-06
2799
- },
2800
- {
2801
- "epoch": 18.64,
2802
- "grad_norm": 0.004608154296875,
2803
- "learning_rate": 6.800000000000001e-07,
2804
- "loss": 0.0,
2805
- "step": 2330
2806
- },
2807
- {
2808
- "epoch": 18.64,
2809
- "step": 2330,
2810
- "train/loss": 5.642308224196313e-06
2811
- },
2812
- {
2813
- "epoch": 18.72,
2814
- "grad_norm": 0.033935546875,
2815
- "learning_rate": 6.4e-07,
2816
- "loss": 0.0,
2817
- "step": 2340
2818
- },
2819
- {
2820
- "epoch": 18.72,
2821
- "step": 2340,
2822
- "train/loss": 2.283810999870184e-06
2823
- },
2824
- {
2825
- "epoch": 18.8,
2826
- "grad_norm": 0.005828857421875,
2827
- "learning_rate": 6.000000000000001e-07,
2828
- "loss": 0.0,
2829
- "step": 2350
2830
- },
2831
- {
2832
- "epoch": 18.8,
2833
- "step": 2350,
2834
- "train/loss": 1.9907624846382532e-06
2835
- },
2836
- {
2837
- "epoch": 18.88,
2838
- "grad_norm": 0.04931640625,
2839
- "learning_rate": 5.6e-07,
2840
- "loss": 0.0,
2841
- "step": 2360
2842
- },
2843
- {
2844
- "epoch": 18.88,
2845
- "step": 2360,
2846
- "train/loss": 1.4662552985100774e-06
2847
- },
2848
- {
2849
- "epoch": 18.96,
2850
- "grad_norm": 0.291015625,
2851
- "learning_rate": 5.2e-07,
2852
- "loss": 0.0001,
2853
- "step": 2370
2854
- },
2855
- {
2856
- "epoch": 18.96,
2857
- "step": 2370,
2858
- "train/loss": 9.493099241808522e-06
2859
- },
2860
- {
2861
- "epoch": 19.04,
2862
- "grad_norm": 0.00116729736328125,
2863
- "learning_rate": 4.800000000000001e-07,
2864
- "loss": 0.0,
2865
- "step": 2380
2866
- },
2867
- {
2868
- "epoch": 19.04,
2869
- "step": 2380,
2870
- "train/loss": 9.761146429809742e-06
2871
- },
2872
- {
2873
- "epoch": 19.12,
2874
- "grad_norm": 0.006927490234375,
2875
- "learning_rate": 4.4e-07,
2876
- "loss": 0.0,
2877
- "step": 2390
2878
- },
2879
- {
2880
- "epoch": 19.12,
2881
- "step": 2390,
2882
- "train/loss": 2.536246756790206e-05
2883
- },
2884
- {
2885
- "epoch": 19.2,
2886
- "grad_norm": 0.004608154296875,
2887
- "learning_rate": 4.0000000000000003e-07,
2888
- "loss": 0.0,
2889
- "step": 2400
2890
- },
2891
- {
2892
- "epoch": 19.2,
2893
- "step": 2400,
2894
- "train/loss": 8.023907867027447e-06
2895
- },
2896
- {
2897
- "epoch": 19.28,
2898
- "grad_norm": 0.0027618408203125,
2899
- "learning_rate": 3.6e-07,
2900
- "loss": 0.0,
2901
- "step": 2410
2902
- },
2903
- {
2904
- "epoch": 19.28,
2905
- "step": 2410,
2906
- "train/loss": 0.0006006373441778123
2907
- },
2908
- {
2909
- "epoch": 19.36,
2910
- "grad_norm": 0.002716064453125,
2911
- "learning_rate": 3.2e-07,
2912
- "loss": 0.0001,
2913
- "step": 2420
2914
- },
2915
- {
2916
- "epoch": 19.36,
2917
- "step": 2420,
2918
- "train/loss": 7.440636977662507e-07
2919
- },
2920
- {
2921
- "epoch": 19.44,
2922
- "grad_norm": 0.001617431640625,
2923
- "learning_rate": 2.8e-07,
2924
- "loss": 0.0,
2925
- "step": 2430
2926
- },
2927
- {
2928
- "epoch": 19.44,
2929
- "step": 2430,
2930
- "train/loss": 1.9689025521074655e-06
2931
- },
2932
- {
2933
- "epoch": 19.52,
2934
- "grad_norm": 0.005767822265625,
2935
- "learning_rate": 2.4000000000000003e-07,
2936
- "loss": 0.0,
2937
- "step": 2440
2938
- },
2939
- {
2940
- "epoch": 19.52,
2941
- "step": 2440,
2942
- "train/loss": 1.3217643754614983e-05
2943
- },
2944
- {
2945
- "epoch": 19.6,
2946
- "grad_norm": 0.018310546875,
2947
- "learning_rate": 2.0000000000000002e-07,
2948
- "loss": 0.0,
2949
- "step": 2450
2950
- },
2951
- {
2952
- "epoch": 19.6,
2953
- "step": 2450,
2954
- "train/loss": 4.347400681581348e-05
2955
- },
2956
- {
2957
- "epoch": 19.68,
2958
- "grad_norm": 0.03955078125,
2959
- "learning_rate": 1.6e-07,
2960
- "loss": 0.0,
2961
- "step": 2460
2962
- },
2963
- {
2964
- "epoch": 19.68,
2965
- "step": 2460,
2966
- "train/loss": 2.077109911624575e-06
2967
- },
2968
- {
2969
- "epoch": 19.76,
2970
- "grad_norm": 0.0034332275390625,
2971
- "learning_rate": 1.2000000000000002e-07,
2972
- "loss": 0.0001,
2973
- "step": 2470
2974
- },
2975
- {
2976
- "epoch": 19.76,
2977
- "step": 2470,
2978
- "train/loss": 1.724543608361273e-06
2979
- },
2980
- {
2981
- "epoch": 19.84,
2982
- "grad_norm": 0.01531982421875,
2983
- "learning_rate": 8e-08,
2984
- "loss": 0.0,
2985
- "step": 2480
2986
- },
2987
- {
2988
- "epoch": 19.84,
2989
- "step": 2480,
2990
- "train/loss": 1.8618636659084586e-06
2991
- },
2992
- {
2993
- "epoch": 19.92,
2994
- "grad_norm": 0.000965118408203125,
2995
- "learning_rate": 4e-08,
2996
- "loss": 0.0,
2997
- "step": 2490
2998
- },
2999
- {
3000
- "epoch": 19.92,
3001
- "step": 2490,
3002
- "train/loss": 9.447310844734602e-07
3003
- },
3004
- {
3005
- "epoch": 20.0,
3006
- "grad_norm": 0.318359375,
3007
- "learning_rate": 0.0,
3008
- "loss": 0.0001,
3009
- "step": 2500
3010
- }
3011
- ],
3012
- "logging_steps": 10,
3013
- "max_steps": 2500,
3014
- "num_input_tokens_seen": 0,
3015
- "num_train_epochs": 20,
3016
- "save_steps": 1500,
3017
- "stateful_callbacks": {
3018
- "TrainerControl": {
3019
- "args": {
3020
- "should_epoch_stop": false,
3021
- "should_evaluate": false,
3022
- "should_log": false,
3023
- "should_save": true,
3024
- "should_training_stop": true
3025
- },
3026
- "attributes": {}
3027
- }
3028
- },
3029
- "total_flos": 0.0,
3030
- "train_batch_size": 4,
3031
- "trial_name": null,
3032
- "trial_params": null
3033
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0bb0bbe398bf6f1c52facbfb16a7e403cb24f51e43c07eb463234278d7eaa04
3
- size 5777
 
 
 
 
sft-clevr-1106-20epoch/checkpoint-2500/vocab.json DELETED
The diff for this file is too large to render. See raw diff