shumi2011 commited on
Commit
8c408f0
·
verified ·
1 Parent(s): f50e1cd

Upload LoRA adapter và processor sau khi train

Browse files
.gitattributes CHANGED
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  checkpoint-402/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  checkpoint-402/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ checkpoint-563/tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -39,7 +39,7 @@ This model was trained with SFT.
39
  ### Framework versions
40
 
41
  - PEFT 0.18.1
42
- - TRL: 0.29.1
43
  - Transformers: 5.4.0
44
  - Pytorch: 2.10.0+cu128
45
  - Datasets: 4.8.4
 
39
  ### Framework versions
40
 
41
  - PEFT 0.18.1
42
+ - TRL: 1.0.0
43
  - Transformers: 5.4.0
44
  - Pytorch: 2.10.0+cu128
45
  - Datasets: 4.8.4
adapter_config.json CHANGED
@@ -30,12 +30,12 @@
30
  "revision": null,
31
  "target_modules": [
32
  "k_proj",
33
- "o_proj",
34
- "down_proj",
35
- "q_proj",
36
  "gate_proj",
 
37
  "up_proj",
38
- "v_proj"
 
 
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
 
30
  "revision": null,
31
  "target_modules": [
32
  "k_proj",
 
 
 
33
  "gate_proj",
34
+ "v_proj",
35
  "up_proj",
36
+ "down_proj",
37
+ "o_proj",
38
+ "q_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43a8ea11e2eceebf2c89454ce0ca1845dcfed115a1cfc69b7d8be2fa14e5c8cc
3
  size 65675408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cb55cb6b8228cc478ad98d9207b252096003053d7d2de477994351d55ef57a0
3
  size 65675408
checkpoint-563/README.md ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/gemma-3-4b-it
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:google/gemma-3-4b-it
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ ---
12
+
13
+ # Model Card for Model ID
14
+
15
+ <!-- Provide a quick summary of what the model is/does. -->
16
+
17
+
18
+
19
+ ## Model Details
20
+
21
+ ### Model Description
22
+
23
+ <!-- Provide a longer summary of what this model is. -->
24
+
25
+
26
+
27
+ - **Developed by:** [More Information Needed]
28
+ - **Funded by [optional]:** [More Information Needed]
29
+ - **Shared by [optional]:** [More Information Needed]
30
+ - **Model type:** [More Information Needed]
31
+ - **Language(s) (NLP):** [More Information Needed]
32
+ - **License:** [More Information Needed]
33
+ - **Finetuned from model [optional]:** [More Information Needed]
34
+
35
+ ### Model Sources [optional]
36
+
37
+ <!-- Provide the basic links for the model. -->
38
+
39
+ - **Repository:** [More Information Needed]
40
+ - **Paper [optional]:** [More Information Needed]
41
+ - **Demo [optional]:** [More Information Needed]
42
+
43
+ ## Uses
44
+
45
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
46
+
47
+ ### Direct Use
48
+
49
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Downstream Use [optional]
54
+
55
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
56
+
57
+ [More Information Needed]
58
+
59
+ ### Out-of-Scope Use
60
+
61
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ## Bias, Risks, and Limitations
66
+
67
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
68
+
69
+ [More Information Needed]
70
+
71
+ ### Recommendations
72
+
73
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
74
+
75
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
76
+
77
+ ## How to Get Started with the Model
78
+
79
+ Use the code below to get started with the model.
80
+
81
+ [More Information Needed]
82
+
83
+ ## Training Details
84
+
85
+ ### Training Data
86
+
87
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
88
+
89
+ [More Information Needed]
90
+
91
+ ### Training Procedure
92
+
93
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
94
+
95
+ #### Preprocessing [optional]
96
+
97
+ [More Information Needed]
98
+
99
+
100
+ #### Training Hyperparameters
101
+
102
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
103
+
104
+ #### Speeds, Sizes, Times [optional]
105
+
106
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
107
+
108
+ [More Information Needed]
109
+
110
+ ## Evaluation
111
+
112
+ <!-- This section describes the evaluation protocols and provides the results. -->
113
+
114
+ ### Testing Data, Factors & Metrics
115
+
116
+ #### Testing Data
117
+
118
+ <!-- This should link to a Dataset Card if possible. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Factors
123
+
124
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
125
+
126
+ [More Information Needed]
127
+
128
+ #### Metrics
129
+
130
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
131
+
132
+ [More Information Needed]
133
+
134
+ ### Results
135
+
136
+ [More Information Needed]
137
+
138
+ #### Summary
139
+
140
+
141
+
142
+ ## Model Examination [optional]
143
+
144
+ <!-- Relevant interpretability work for the model goes here -->
145
+
146
+ [More Information Needed]
147
+
148
+ ## Environmental Impact
149
+
150
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
151
+
152
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
153
+
154
+ - **Hardware Type:** [More Information Needed]
155
+ - **Hours used:** [More Information Needed]
156
+ - **Cloud Provider:** [More Information Needed]
157
+ - **Compute Region:** [More Information Needed]
158
+ - **Carbon Emitted:** [More Information Needed]
159
+
160
+ ## Technical Specifications [optional]
161
+
162
+ ### Model Architecture and Objective
163
+
164
+ [More Information Needed]
165
+
166
+ ### Compute Infrastructure
167
+
168
+ [More Information Needed]
169
+
170
+ #### Hardware
171
+
172
+ [More Information Needed]
173
+
174
+ #### Software
175
+
176
+ [More Information Needed]
177
+
178
+ ## Citation [optional]
179
+
180
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
181
+
182
+ **BibTeX:**
183
+
184
+ [More Information Needed]
185
+
186
+ **APA:**
187
+
188
+ [More Information Needed]
189
+
190
+ ## Glossary [optional]
191
+
192
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
193
+
194
+ [More Information Needed]
195
+
196
+ ## More Information [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Authors [optional]
201
+
202
+ [More Information Needed]
203
+
204
+ ## Model Card Contact
205
+
206
+ [More Information Needed]
207
+ ### Framework versions
208
+
209
+ - PEFT 0.18.1
checkpoint-563/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "google/gemma-3-4b-it",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": null,
25
+ "peft_type": "LORA",
26
+ "peft_version": "0.18.1",
27
+ "qalora_group_size": 16,
28
+ "r": 16,
29
+ "rank_pattern": {},
30
+ "revision": null,
31
+ "target_modules": [
32
+ "k_proj",
33
+ "gate_proj",
34
+ "v_proj",
35
+ "up_proj",
36
+ "down_proj",
37
+ "o_proj",
38
+ "q_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
checkpoint-563/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cb55cb6b8228cc478ad98d9207b252096003053d7d2de477994351d55ef57a0
3
+ size 65675408
checkpoint-563/chat_template.jinja ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {{ bos_token }}
2
+ {%- if messages[0]['role'] == 'system' -%}
3
+ {%- if messages[0]['content'] is string -%}
4
+ {%- set first_user_prefix = messages[0]['content'] + '
5
+
6
+ ' -%}
7
+ {%- else -%}
8
+ {%- set first_user_prefix = messages[0]['content'][0]['text'] + '
9
+
10
+ ' -%}
11
+ {%- endif -%}
12
+ {%- set loop_messages = messages[1:] -%}
13
+ {%- else -%}
14
+ {%- set first_user_prefix = "" -%}
15
+ {%- set loop_messages = messages -%}
16
+ {%- endif -%}
17
+ {%- for message in loop_messages -%}
18
+ {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}
19
+ {{ raise_exception("Conversation roles must alternate user/assistant/user/assistant/...") }}
20
+ {%- endif -%}
21
+ {%- if (message['role'] == 'assistant') -%}
22
+ {%- set role = "model" -%}
23
+ {%- else -%}
24
+ {%- set role = message['role'] -%}
25
+ {%- endif -%}
26
+ {{ '<start_of_turn>' + role + '
27
+ ' + (first_user_prefix if loop.first else "") }}
28
+ {%- if message['content'] is string -%}
29
+ {{ message['content'] | trim }}
30
+ {%- elif message['content'] is iterable -%}
31
+ {%- for item in message['content'] -%}
32
+ {%- if item['type'] == 'image' -%}
33
+ {{ '<start_of_image>' }}
34
+ {%- elif item['type'] == 'text' -%}
35
+ {{ item['text'] | trim }}
36
+ {%- endif -%}
37
+ {%- endfor -%}
38
+ {%- else -%}
39
+ {{ raise_exception("Invalid content type") }}
40
+ {%- endif -%}
41
+ {{ '<end_of_turn>
42
+ ' }}
43
+ {%- endfor -%}
44
+ {%- if add_generation_prompt -%}
45
+ {{'<start_of_turn>model
46
+ '}}
47
+ {%- endif -%}
checkpoint-563/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:147ecdea7929fde43e244c73dd544abb2ed359ded440d922e8dee6f5dc3009fe
3
+ size 119618359
checkpoint-563/processor_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "image_processor": {
3
+ "do_convert_rgb": null,
4
+ "do_normalize": true,
5
+ "do_rescale": true,
6
+ "do_resize": true,
7
+ "image_mean": [
8
+ 0.5,
9
+ 0.5,
10
+ 0.5
11
+ ],
12
+ "image_processor_type": "Gemma3ImageProcessor",
13
+ "image_seq_length": 256,
14
+ "image_std": [
15
+ 0.5,
16
+ 0.5,
17
+ 0.5
18
+ ],
19
+ "resample": 2,
20
+ "rescale_factor": 0.00392156862745098,
21
+ "size": {
22
+ "height": 896,
23
+ "width": 896
24
+ }
25
+ },
26
+ "image_seq_length": 256,
27
+ "processor_class": "Gemma3Processor"
28
+ }
checkpoint-563/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3cd3621b8b24d40d23ff41931c6ba8079e1bd59ff5085e42505384d72c06a13
3
+ size 14709
checkpoint-563/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fae9377d103c0880629be5550a0dfd948a5c6f4777668a5c22004bc9daed5d9f
3
+ size 1465
checkpoint-563/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daab2354f8a74e70d70b4d1f804939b68a8c9624dd06cb7858e52dd8970e9726
3
+ size 33384567
checkpoint-563/tokenizer_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "boi_token": "<start_of_image>",
4
+ "bos_token": "<bos>",
5
+ "clean_up_tokenization_spaces": false,
6
+ "eoi_token": "<end_of_image>",
7
+ "eos_token": "<eos>",
8
+ "image_token": "<image_soft_token>",
9
+ "is_local": false,
10
+ "mask_token": "<mask>",
11
+ "model_max_length": 1000000000000000019884624838656,
12
+ "model_specific_special_tokens": {
13
+ "boi_token": "<start_of_image>",
14
+ "eoi_token": "<end_of_image>",
15
+ "image_token": "<image_soft_token>"
16
+ },
17
+ "pad_token": "<pad>",
18
+ "processor_class": "Gemma3Processor",
19
+ "sp_model_kwargs": null,
20
+ "spaces_between_special_tokens": false,
21
+ "tokenizer_class": "GemmaTokenizer",
22
+ "unk_token": "<unk>",
23
+ "use_default_system_prompt": false
24
+ }
checkpoint-563/trainer_state.json ADDED
@@ -0,0 +1,605 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 563,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "entropy": 1.2997896295040845,
14
+ "epoch": 0.017777777777777778,
15
+ "grad_norm": 1.3046875,
16
+ "learning_rate": 1.6071428571428572e-05,
17
+ "loss": 5.099749374389648,
18
+ "mean_token_accuracy": 0.4248364338651299,
19
+ "num_tokens": 15791.0,
20
+ "step": 10
21
+ },
22
+ {
23
+ "entropy": 1.6972164057195187,
24
+ "epoch": 0.035555555555555556,
25
+ "grad_norm": 0.333984375,
26
+ "learning_rate": 3.392857142857143e-05,
27
+ "loss": 4.261620330810547,
28
+ "mean_token_accuracy": 0.42575850784778596,
29
+ "num_tokens": 30788.0,
30
+ "step": 20
31
+ },
32
+ {
33
+ "entropy": 2.371405544877052,
34
+ "epoch": 0.05333333333333334,
35
+ "grad_norm": 0.2451171875,
36
+ "learning_rate": 5.1785714285714296e-05,
37
+ "loss": 3.410958480834961,
38
+ "mean_token_accuracy": 0.454820466786623,
39
+ "num_tokens": 45784.0,
40
+ "step": 30
41
+ },
42
+ {
43
+ "entropy": 2.792352019995451,
44
+ "epoch": 0.07111111111111111,
45
+ "grad_norm": 0.388671875,
46
+ "learning_rate": 6.964285714285715e-05,
47
+ "loss": 2.760978126525879,
48
+ "mean_token_accuracy": 0.5088610924780369,
49
+ "num_tokens": 62673.0,
50
+ "step": 40
51
+ },
52
+ {
53
+ "entropy": 2.2824966341257094,
54
+ "epoch": 0.08888888888888889,
55
+ "grad_norm": 0.35546875,
56
+ "learning_rate": 8.75e-05,
57
+ "loss": 2.191845512390137,
58
+ "mean_token_accuracy": 0.5701939344406128,
59
+ "num_tokens": 76957.0,
60
+ "step": 50
61
+ },
62
+ {
63
+ "entropy": 1.8100679598748683,
64
+ "epoch": 0.10666666666666667,
65
+ "grad_norm": 0.31640625,
66
+ "learning_rate": 9.999136119166803e-05,
67
+ "loss": 1.7198040008544921,
68
+ "mean_token_accuracy": 0.650282097607851,
69
+ "num_tokens": 90947.0,
70
+ "step": 60
71
+ },
72
+ {
73
+ "entropy": 1.237974463403225,
74
+ "epoch": 0.12444444444444444,
75
+ "grad_norm": 0.337890625,
76
+ "learning_rate": 9.983786540671051e-05,
77
+ "loss": 1.1844627380371093,
78
+ "mean_token_accuracy": 0.7430308632552624,
79
+ "num_tokens": 105311.0,
80
+ "step": 70
81
+ },
82
+ {
83
+ "entropy": 0.8024648647755385,
84
+ "epoch": 0.14222222222222222,
85
+ "grad_norm": 0.466796875,
86
+ "learning_rate": 9.949307432339625e-05,
87
+ "loss": 0.8033592224121093,
88
+ "mean_token_accuracy": 0.8211114652454853,
89
+ "num_tokens": 120739.0,
90
+ "step": 80
91
+ },
92
+ {
93
+ "entropy": 0.6481792986392975,
94
+ "epoch": 0.16,
95
+ "grad_norm": 0.3515625,
96
+ "learning_rate": 9.895831137146318e-05,
97
+ "loss": 0.6663334846496582,
98
+ "mean_token_accuracy": 0.8503929443657399,
99
+ "num_tokens": 135550.0,
100
+ "step": 90
101
+ },
102
+ {
103
+ "entropy": 0.5209066523239017,
104
+ "epoch": 0.17777777777777778,
105
+ "grad_norm": 0.220703125,
106
+ "learning_rate": 9.82356291596578e-05,
107
+ "loss": 0.5501353740692139,
108
+ "mean_token_accuracy": 0.8712642557919026,
109
+ "num_tokens": 150124.0,
110
+ "step": 100
111
+ },
112
+ {
113
+ "entropy": 0.4373117024078965,
114
+ "epoch": 0.19555555555555557,
115
+ "grad_norm": 0.267578125,
116
+ "learning_rate": 9.732780159709912e-05,
117
+ "loss": 0.3997534513473511,
118
+ "mean_token_accuracy": 0.9081158190965652,
119
+ "num_tokens": 164899.0,
120
+ "step": 110
121
+ },
122
+ {
123
+ "entropy": 0.36014524921774865,
124
+ "epoch": 0.21333333333333335,
125
+ "grad_norm": 0.185546875,
126
+ "learning_rate": 9.623831324603754e-05,
127
+ "loss": 0.3664620161056519,
128
+ "mean_token_accuracy": 0.9099370762705803,
129
+ "num_tokens": 180128.0,
130
+ "step": 120
131
+ },
132
+ {
133
+ "entropy": 0.29999935105443,
134
+ "epoch": 0.2311111111111111,
135
+ "grad_norm": 0.255859375,
136
+ "learning_rate": 9.497134594687634e-05,
137
+ "loss": 0.2984978437423706,
138
+ "mean_token_accuracy": 0.9242603577673435,
139
+ "num_tokens": 195976.0,
140
+ "step": 130
141
+ },
142
+ {
143
+ "entropy": 0.31300447108224033,
144
+ "epoch": 0.24888888888888888,
145
+ "grad_norm": 0.2060546875,
146
+ "learning_rate": 9.353176276679396e-05,
147
+ "loss": 0.32675857543945314,
148
+ "mean_token_accuracy": 0.9216666355729103,
149
+ "num_tokens": 211055.0,
150
+ "step": 140
151
+ },
152
+ {
153
+ "entropy": 0.2710464348085225,
154
+ "epoch": 0.26666666666666666,
155
+ "grad_norm": 0.2578125,
156
+ "learning_rate": 9.192508933357753e-05,
157
+ "loss": 0.2727458715438843,
158
+ "mean_token_accuracy": 0.9310060679912567,
159
+ "num_tokens": 227341.0,
160
+ "step": 150
161
+ },
162
+ {
163
+ "entropy": 0.30849818857386707,
164
+ "epoch": 0.28444444444444444,
165
+ "grad_norm": 0.271484375,
166
+ "learning_rate": 9.015749262631536e-05,
167
+ "loss": 0.30527050495147706,
168
+ "mean_token_accuracy": 0.9228560633957386,
169
+ "num_tokens": 242924.0,
170
+ "step": 160
171
+ },
172
+ {
173
+ "entropy": 0.282581620849669,
174
+ "epoch": 0.3022222222222222,
175
+ "grad_norm": 0.3828125,
176
+ "learning_rate": 8.823575730435693e-05,
177
+ "loss": 0.2990081548690796,
178
+ "mean_token_accuracy": 0.9226435236632824,
179
+ "num_tokens": 256777.0,
180
+ "step": 170
181
+ },
182
+ {
183
+ "entropy": 0.2864753663539886,
184
+ "epoch": 0.32,
185
+ "grad_norm": 0.35546875,
186
+ "learning_rate": 8.616725966539832e-05,
187
+ "loss": 0.3073178768157959,
188
+ "mean_token_accuracy": 0.9258665904402733,
189
+ "num_tokens": 272913.0,
190
+ "step": 180
191
+ },
192
+ {
193
+ "entropy": 0.26701974822208285,
194
+ "epoch": 0.3377777777777778,
195
+ "grad_norm": 0.2578125,
196
+ "learning_rate": 8.395993933265101e-05,
197
+ "loss": 0.261165452003479,
198
+ "mean_token_accuracy": 0.9338324561715126,
199
+ "num_tokens": 287769.0,
200
+ "step": 190
201
+ },
202
+ {
203
+ "entropy": 0.23425031434744598,
204
+ "epoch": 0.35555555555555557,
205
+ "grad_norm": 0.2373046875,
206
+ "learning_rate": 8.162226877976887e-05,
207
+ "loss": 0.21241703033447265,
208
+ "mean_token_accuracy": 0.9439801961183548,
209
+ "num_tokens": 303589.0,
210
+ "step": 200
211
+ },
212
+ {
213
+ "entropy": 0.26449646847322583,
214
+ "epoch": 0.37333333333333335,
215
+ "grad_norm": 0.271484375,
216
+ "learning_rate": 7.916322081050709e-05,
217
+ "loss": 0.27304508686065676,
218
+ "mean_token_accuracy": 0.9296576961874962,
219
+ "num_tokens": 317660.0,
220
+ "step": 210
221
+ },
222
+ {
223
+ "entropy": 0.15648896424099804,
224
+ "epoch": 0.39111111111111113,
225
+ "grad_norm": 0.1826171875,
226
+ "learning_rate": 7.659223411793798e-05,
227
+ "loss": 0.15494911670684813,
228
+ "mean_token_accuracy": 0.9541799262166023,
229
+ "num_tokens": 333271.0,
230
+ "step": 220
231
+ },
232
+ {
233
+ "entropy": 0.21079173348844052,
234
+ "epoch": 0.4088888888888889,
235
+ "grad_norm": 0.283203125,
236
+ "learning_rate": 7.391917705541927e-05,
237
+ "loss": 0.22306714057922364,
238
+ "mean_token_accuracy": 0.9384914793074131,
239
+ "num_tokens": 347985.0,
240
+ "step": 230
241
+ },
242
+ {
243
+ "entropy": 0.2048337606713176,
244
+ "epoch": 0.4266666666666667,
245
+ "grad_norm": 0.126953125,
246
+ "learning_rate": 7.115430975837457e-05,
247
+ "loss": 0.19801312685012817,
248
+ "mean_token_accuracy": 0.9438701763749122,
249
+ "num_tokens": 363032.0,
250
+ "step": 240
251
+ },
252
+ {
253
+ "entropy": 0.22145606707781554,
254
+ "epoch": 0.4444444444444444,
255
+ "grad_norm": 0.115234375,
256
+ "learning_rate": 6.830824476227646e-05,
257
+ "loss": 0.21796202659606934,
258
+ "mean_token_accuracy": 0.9396323539316654,
259
+ "num_tokens": 377070.0,
260
+ "step": 250
261
+ },
262
+ {
263
+ "entropy": 0.22976100631058216,
264
+ "epoch": 0.4622222222222222,
265
+ "grad_norm": 0.21484375,
266
+ "learning_rate": 6.539190626799366e-05,
267
+ "loss": 0.21554169654846192,
268
+ "mean_token_accuracy": 0.938993276655674,
269
+ "num_tokens": 392460.0,
270
+ "step": 260
271
+ },
272
+ {
273
+ "entropy": 0.16580691728740932,
274
+ "epoch": 0.48,
275
+ "grad_norm": 0.1298828125,
276
+ "learning_rate": 6.241648821085666e-05,
277
+ "loss": 0.16170507669448853,
278
+ "mean_token_accuracy": 0.9502397567033768,
279
+ "num_tokens": 408721.0,
280
+ "step": 270
281
+ },
282
+ {
283
+ "entropy": 0.17671420807018876,
284
+ "epoch": 0.49777777777777776,
285
+ "grad_norm": 0.2890625,
286
+ "learning_rate": 5.939341129438739e-05,
287
+ "loss": 0.19148651361465455,
288
+ "mean_token_accuracy": 0.9441343322396278,
289
+ "num_tokens": 423511.0,
290
+ "step": 280
291
+ },
292
+ {
293
+ "entropy": 0.19226484168320895,
294
+ "epoch": 0.5155555555555555,
295
+ "grad_norm": 0.2021484375,
296
+ "learning_rate": 5.633427915361261e-05,
297
+ "loss": 0.1809281587600708,
298
+ "mean_token_accuracy": 0.9457803666591644,
299
+ "num_tokens": 438350.0,
300
+ "step": 290
301
+ },
302
+ {
303
+ "entropy": 0.1846911649219692,
304
+ "epoch": 0.5333333333333333,
305
+ "grad_norm": 0.10791015625,
306
+ "learning_rate": 5.325083381622165e-05,
307
+ "loss": 0.19349638223648072,
308
+ "mean_token_accuracy": 0.9445783741772175,
309
+ "num_tokens": 454504.0,
310
+ "step": 300
311
+ },
312
+ {
313
+ "entropy": 0.19011163590475916,
314
+ "epoch": 0.5511111111111111,
315
+ "grad_norm": 0.146484375,
316
+ "learning_rate": 5.01549106325243e-05,
317
+ "loss": 0.1816372752189636,
318
+ "mean_token_accuracy": 0.9470942720770836,
319
+ "num_tokens": 469888.0,
320
+ "step": 310
321
+ },
322
+ {
323
+ "entropy": 0.21089051999151706,
324
+ "epoch": 0.5688888888888889,
325
+ "grad_norm": 0.142578125,
326
+ "learning_rate": 4.705839284720376e-05,
327
+ "loss": 0.20321893692016602,
328
+ "mean_token_accuracy": 0.9423739515244961,
329
+ "num_tokens": 484515.0,
330
+ "step": 320
331
+ },
332
+ {
333
+ "entropy": 0.15668439203873277,
334
+ "epoch": 0.5866666666666667,
335
+ "grad_norm": 0.099609375,
336
+ "learning_rate": 4.397316598723385e-05,
337
+ "loss": 0.152730131149292,
338
+ "mean_token_accuracy": 0.954561373591423,
339
+ "num_tokens": 500366.0,
340
+ "step": 330
341
+ },
342
+ {
343
+ "entropy": 0.19265495147556067,
344
+ "epoch": 0.6044444444444445,
345
+ "grad_norm": 0.2490234375,
346
+ "learning_rate": 4.0911072241036194e-05,
347
+ "loss": 0.18428765535354613,
348
+ "mean_token_accuracy": 0.9454522147774697,
349
+ "num_tokens": 514445.0,
350
+ "step": 340
351
+ },
352
+ {
353
+ "entropy": 0.14791145911440254,
354
+ "epoch": 0.6222222222222222,
355
+ "grad_norm": 0.126953125,
356
+ "learning_rate": 3.788386500398583e-05,
357
+ "loss": 0.15320039987564088,
358
+ "mean_token_accuracy": 0.9516465291380882,
359
+ "num_tokens": 529805.0,
360
+ "step": 350
361
+ },
362
+ {
363
+ "entropy": 0.14857742255553602,
364
+ "epoch": 0.64,
365
+ "grad_norm": 0.09033203125,
366
+ "learning_rate": 3.49031637647361e-05,
367
+ "loss": 0.1594814896583557,
368
+ "mean_token_accuracy": 0.9535981945693492,
369
+ "num_tokens": 545883.0,
370
+ "step": 360
371
+ },
372
+ {
373
+ "entropy": 0.2208730909973383,
374
+ "epoch": 0.6577777777777778,
375
+ "grad_norm": 0.08837890625,
376
+ "learning_rate": 3.1980409505524544e-05,
377
+ "loss": 0.19757508039474486,
378
+ "mean_token_accuracy": 0.9436426095664501,
379
+ "num_tokens": 559836.0,
380
+ "step": 370
381
+ },
382
+ {
383
+ "entropy": 0.1493451208807528,
384
+ "epoch": 0.6755555555555556,
385
+ "grad_norm": 0.12109375,
386
+ "learning_rate": 2.91268207876494e-05,
387
+ "loss": 0.15581512451171875,
388
+ "mean_token_accuracy": 0.9540682502090931,
389
+ "num_tokens": 575627.0,
390
+ "step": 380
391
+ },
392
+ {
393
+ "entropy": 0.13959282217547297,
394
+ "epoch": 0.6933333333333334,
395
+ "grad_norm": 0.1533203125,
396
+ "learning_rate": 2.635335069067617e-05,
397
+ "loss": 0.12903414964675902,
398
+ "mean_token_accuracy": 0.9592652179300785,
399
+ "num_tokens": 590314.0,
400
+ "step": 390
401
+ },
402
+ {
403
+ "entropy": 0.16119700381532312,
404
+ "epoch": 0.7111111111111111,
405
+ "grad_norm": 0.1318359375,
406
+ "learning_rate": 2.367064477065652e-05,
407
+ "loss": 0.14327847957611084,
408
+ "mean_token_accuracy": 0.9542674884200096,
409
+ "num_tokens": 605721.0,
410
+ "step": 400
411
+ },
412
+ {
413
+ "entropy": 0.151368910074234,
414
+ "epoch": 0.7288888888888889,
415
+ "grad_norm": 0.1484375,
416
+ "learning_rate": 2.108900019873103e-05,
417
+ "loss": 0.14610207080841064,
418
+ "mean_token_accuracy": 0.9552296213805676,
419
+ "num_tokens": 621020.0,
420
+ "step": 410
421
+ },
422
+ {
423
+ "entropy": 0.14151866482570769,
424
+ "epoch": 0.7466666666666667,
425
+ "grad_norm": 0.11328125,
426
+ "learning_rate": 1.8618326236955907e-05,
427
+ "loss": 0.13566198348999023,
428
+ "mean_token_accuracy": 0.9568316303193569,
429
+ "num_tokens": 636597.0,
430
+ "step": 420
431
+ },
432
+ {
433
+ "entropy": 0.1339370148256421,
434
+ "epoch": 0.7644444444444445,
435
+ "grad_norm": 0.1689453125,
436
+ "learning_rate": 1.626810620306163e-05,
437
+ "loss": 0.1301966667175293,
438
+ "mean_token_accuracy": 0.957643074542284,
439
+ "num_tokens": 652071.0,
440
+ "step": 430
441
+ },
442
+ {
443
+ "entropy": 0.15190135017037393,
444
+ "epoch": 0.7822222222222223,
445
+ "grad_norm": 0.09130859375,
446
+ "learning_rate": 1.4047361070135995e-05,
447
+ "loss": 0.1389269709587097,
448
+ "mean_token_accuracy": 0.9543100669980049,
449
+ "num_tokens": 668341.0,
450
+ "step": 440
451
+ },
452
+ {
453
+ "entropy": 0.18330826219171287,
454
+ "epoch": 0.8,
455
+ "grad_norm": 0.24609375,
456
+ "learning_rate": 1.1964614840949002e-05,
457
+ "loss": 0.18672417402267455,
458
+ "mean_token_accuracy": 0.9420441940426827,
459
+ "num_tokens": 683688.0,
460
+ "step": 450
461
+ },
462
+ {
463
+ "entropy": 0.17232957119122147,
464
+ "epoch": 0.8177777777777778,
465
+ "grad_norm": 0.1044921875,
466
+ "learning_rate": 1.0027861829824952e-05,
467
+ "loss": 0.16492395401000975,
468
+ "mean_token_accuracy": 0.9488118067383766,
469
+ "num_tokens": 699667.0,
470
+ "step": 460
471
+ },
472
+ {
473
+ "entropy": 0.15511126685887575,
474
+ "epoch": 0.8355555555555556,
475
+ "grad_norm": 0.11328125,
476
+ "learning_rate": 8.244535977645585e-06,
477
+ "loss": 0.15686993598937987,
478
+ "mean_token_accuracy": 0.9512658596038819,
479
+ "num_tokens": 713516.0,
480
+ "step": 470
481
+ },
482
+ {
483
+ "entropy": 0.1436025496572256,
484
+ "epoch": 0.8533333333333334,
485
+ "grad_norm": 0.0927734375,
486
+ "learning_rate": 6.621482317764105e-06,
487
+ "loss": 0.13364784717559813,
488
+ "mean_token_accuracy": 0.9571717426180839,
489
+ "num_tokens": 729126.0,
490
+ "step": 480
491
+ },
492
+ {
493
+ "entropy": 0.1506779951043427,
494
+ "epoch": 0.8711111111111111,
495
+ "grad_norm": 0.1611328125,
496
+ "learning_rate": 5.164930702353782e-06,
497
+ "loss": 0.14193692207336425,
498
+ "mean_token_accuracy": 0.9549389965832233,
499
+ "num_tokens": 744962.0,
500
+ "step": 490
501
+ },
502
+ {
503
+ "entropy": 0.1640948430635035,
504
+ "epoch": 0.8888888888888888,
505
+ "grad_norm": 0.3046875,
506
+ "learning_rate": 3.880471890038967e-06,
507
+ "loss": 0.16054811477661132,
508
+ "mean_token_accuracy": 0.9514626495540142,
509
+ "num_tokens": 759678.0,
510
+ "step": 500
511
+ },
512
+ {
513
+ "entropy": 0.16496095675975084,
514
+ "epoch": 0.9066666666666666,
515
+ "grad_norm": 0.234375,
516
+ "learning_rate": 2.7730360865923956e-06,
517
+ "loss": 0.16119725704193116,
518
+ "mean_token_accuracy": 0.9486441940069199,
519
+ "num_tokens": 774383.0,
520
+ "step": 510
521
+ },
522
+ {
523
+ "entropy": 0.17073705019429325,
524
+ "epoch": 0.9244444444444444,
525
+ "grad_norm": 0.12353515625,
526
+ "learning_rate": 1.8468740210672076e-06,
527
+ "loss": 0.1567433476448059,
528
+ "mean_token_accuracy": 0.950737326592207,
529
+ "num_tokens": 788196.0,
530
+ "step": 520
531
+ },
532
+ {
533
+ "entropy": 0.14311794554814697,
534
+ "epoch": 0.9422222222222222,
535
+ "grad_norm": 0.2431640625,
536
+ "learning_rate": 1.1055406300002347e-06,
537
+ "loss": 0.13767447471618652,
538
+ "mean_token_accuracy": 0.9572608590126037,
539
+ "num_tokens": 803510.0,
540
+ "step": 530
541
+ },
542
+ {
543
+ "entropy": 0.1503830960020423,
544
+ "epoch": 0.96,
545
+ "grad_norm": 0.1474609375,
546
+ "learning_rate": 5.518814123121885e-07,
547
+ "loss": 0.13956043720245362,
548
+ "mean_token_accuracy": 0.95562659278512,
549
+ "num_tokens": 819096.0,
550
+ "step": 540
551
+ },
552
+ {
553
+ "entropy": 0.14491902142763138,
554
+ "epoch": 0.9777777777777777,
555
+ "grad_norm": 0.134765625,
556
+ "learning_rate": 1.8802150727962876e-07,
557
+ "loss": 0.13078807592391967,
558
+ "mean_token_accuracy": 0.9562703162431717,
559
+ "num_tokens": 834121.0,
560
+ "step": 550
561
+ },
562
+ {
563
+ "entropy": 0.1624174129217863,
564
+ "epoch": 0.9955555555555555,
565
+ "grad_norm": 0.0888671875,
566
+ "learning_rate": 1.5357537501159423e-08,
567
+ "loss": 0.15792603492736818,
568
+ "mean_token_accuracy": 0.9509337961673736,
569
+ "num_tokens": 850350.0,
570
+ "step": 560
571
+ },
572
+ {
573
+ "epoch": 1.0,
574
+ "eval_entropy": 0.17035142924636604,
575
+ "eval_loss": 0.16859714686870575,
576
+ "eval_mean_token_accuracy": 0.9493516847491265,
577
+ "eval_num_tokens": 854418.0,
578
+ "eval_runtime": 193.5414,
579
+ "eval_samples_per_second": 1.033,
580
+ "eval_steps_per_second": 1.033,
581
+ "step": 563
582
+ }
583
+ ],
584
+ "logging_steps": 10,
585
+ "max_steps": 563,
586
+ "num_input_tokens_seen": 0,
587
+ "num_train_epochs": 1,
588
+ "save_steps": 500,
589
+ "stateful_callbacks": {
590
+ "TrainerControl": {
591
+ "args": {
592
+ "should_epoch_stop": false,
593
+ "should_evaluate": false,
594
+ "should_log": false,
595
+ "should_save": true,
596
+ "should_training_stop": true
597
+ },
598
+ "attributes": {}
599
+ }
600
+ },
601
+ "total_flos": 1.874711111049504e+16,
602
+ "train_batch_size": 1,
603
+ "trial_name": null,
604
+ "trial_params": null
605
+ }
checkpoint-563/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc262f4d0e8c93b76e95a23291278880e2bee6bb924cab03eabbb9c30033301b
3
+ size 5713
runs/Mar31_04-20-31_242ca503c882/events.out.tfevents.1774930831.242ca503c882.3260.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcd800807729c96289422c048ba7b5103d581de1873ce792424d2d297f50aa3b
3
+ size 10153
runs/Mar31_06-15-17_242ca503c882/events.out.tfevents.1774937717.242ca503c882.32504.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8912bd32da89f1d341f20edb5d40cfe81095db33a607e350bed88e7080fbb2de
3
+ size 29629
tokenizer_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "boi_token": "<start_of_image>",
4
+ "bos_token": "<bos>",
5
+ "clean_up_tokenization_spaces": false,
6
+ "eoi_token": "<end_of_image>",
7
+ "eos_token": "<eos>",
8
+ "image_token": "<image_soft_token>",
9
+ "is_local": false,
10
+ "mask_token": "<mask>",
11
+ "model_max_length": 1000000000000000019884624838656,
12
+ "model_specific_special_tokens": {
13
+ "boi_token": "<start_of_image>",
14
+ "eoi_token": "<end_of_image>",
15
+ "image_token": "<image_soft_token>"
16
+ },
17
+ "pad_token": "<pad>",
18
+ "processor_class": "Gemma3Processor",
19
+ "sp_model_kwargs": null,
20
+ "spaces_between_special_tokens": false,
21
+ "tokenizer_class": "GemmaTokenizer",
22
+ "unk_token": "<unk>",
23
+ "use_default_system_prompt": false
24
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ddb27f7a670f95fa93dcb5c98c5d06a951526c8df56cf5945c5a6066148a8206
3
- size 5649
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc262f4d0e8c93b76e95a23291278880e2bee6bb924cab03eabbb9c30033301b
3
+ size 5713