AiAF commited on
Commit
dfaa246
Β·
verified Β·
1 Parent(s): 283c16a

End of training

Browse files
Files changed (2) hide show
  1. README.md +237 -3
  2. debug.log +193 -1
README.md CHANGED
@@ -1,13 +1,247 @@
1
  ---
2
- base_model: google/gemma-2-2b-it
3
  library_name: peft
4
- pipeline_tag: text-generation
 
5
  tags:
6
  - axolotl
7
  - base_model:adapter:google/gemma-2-2b-it
8
  - lora
9
  - transformers
 
 
 
 
 
 
10
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ### Framework versions
12
 
13
- - PEFT 0.17.1
 
 
 
 
 
1
  ---
 
2
  library_name: peft
3
+ license: gemma
4
+ base_model: google/gemma-2-2b-it
5
  tags:
6
  - axolotl
7
  - base_model:adapter:google/gemma-2-2b-it
8
  - lora
9
  - transformers
10
+ datasets:
11
+ - AiAF/conversations
12
+ pipeline_tag: text-generation
13
+ model-index:
14
+ - name: rp-2b
15
+ results: []
16
  ---
17
+
18
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
19
+ should probably proofread and complete it, then remove this comment. -->
20
+
21
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
22
+ <details><summary>See axolotl config</summary>
23
+
24
+ axolotl version: `0.13.0.dev0`
25
+ ```yaml
26
+ # 1. Base Model & Tokenizer
27
+ base_model: google/gemma-2-2b-it
28
+ model_type: AutoModelForCausalLM # Corrected from 'type_of_model' for axolotl
29
+ tokenizer_type: AutoTokenizer
30
+ hub_model_id: AiAF/rp-2b # New model ID for this finetune
31
+ hub_strategy: checkpoint
32
+
33
+ # 2. LoRA / QLoRA Configuration
34
+ load_in_4bit: true
35
+ adapter: qlora
36
+ lora_r: 64
37
+ lora_alpha: 128
38
+ lora_dropout: 0.05
39
+ lora_target_linear: true
40
+
41
+ # 3. Dataset Configuration (TRAIN = streamed)
42
+ streaming: true
43
+ streaming_multipack_buffer_size: 10000
44
+ sample_packing: true
45
+ datasets:
46
+ - path: AiAF/conversations
47
+ data_files: conversations_V3.jsonl
48
+ # revision:
49
+ type: chat_template
50
+ split: train
51
+ field_messages: conversations
52
+ message_property_mappings:
53
+ role: from
54
+ content: value
55
+ chat_template: jinja
56
+ chat_template_jinja: |
57
+ {{ bos_token }}
58
+ {% for m in messages %}
59
+ {% set role = 'model' if m['role']=='assistant' else 'user' %}
60
+ {{ '<start_of_turn>' + role + '\n' + m['content'] | trim + '<end_of_turn>\n' }}
61
+ {% endfor %}
62
+ {% if add_generation_prompt %}
63
+ {{ '<start_of_turn>model\n' }}
64
+ {% endif %}
65
+
66
+ # chat_template_jinja: |
67
+ # {{ bos_token }}
68
+ # {% set last = None %}
69
+ # {% for m in messages %}
70
+ # {% set raw_role = 'model' if m['role']=='assistant' else m['role'] %}
71
+ # {% set role = 'user' if raw_role=='system' else raw_role %}
72
+ # {% if role == last and role == 'user' %}
73
+ # {{ m['content'] | trim }}
74
+ # {% else %}
75
+ # {{ '<start_of_turn>' + role + '\n' + m['content'] | trim + '<end_of_turn>\n' }}
76
+ # {% endif %}
77
+ # {% set last = role %}
78
+ # {% endfor %}
79
+ # {% if add_generation_prompt %}
80
+ # {{ '<start_of_turn>model\n' }}
81
+ # {% endif %}
82
+ roles_to_train: ["assistant"]
83
+ train_on_eos: "turn"
84
+ # Use a fixed (non-streamed) eval file with the same schema/Jinja
85
+ test_datasets:
86
+ - path: .
87
+ name: json
88
+ type: chat_template
89
+ data_files: eval-datasets/shuf-1000_conversations_V2.jsonl # small, representative eval slice
90
+ split: train
91
+ field_messages: conversations
92
+ message_property_mappings:
93
+ role: from
94
+ content: value
95
+ chat_template: jinja
96
+ chat_template_jinja: |
97
+ {{ bos_token }}
98
+ {% for m in messages %}
99
+ {% set role = 'model' if m['role']=='assistant' else 'user' %}
100
+ {{ '<start_of_turn>' + role + '\n' + m['content'] | trim + '<end_of_turn>\n' }}
101
+ {% endfor %}
102
+ {% if add_generation_prompt %}
103
+ {{ '<start_of_turn>model\n' }}
104
+ {% endif %}
105
+ # chat_template_jinja: |
106
+ # {{ bos_token }}
107
+ # {% set last = None %}
108
+ # {% for m in messages %}
109
+ # {% set raw_role = 'model' if m['role']=='assistant' else m['role'] %}
110
+ # {% set role = 'user' if raw_role=='system' else raw_role %}
111
+ # {% if role == last and role == 'user' %}
112
+ # {{ m['content'] | trim }}
113
+ # {% else %}
114
+ # {{ '<start_of_turn>' + role + '\n' + m['content'] | trim + '<end_of_turn>\n' }}
115
+ # {% endif %}
116
+ # {% set last = role %}
117
+ # {% endfor %}
118
+ # {% if add_generation_prompt %}
119
+ # {{ '<start_of_turn>model\n' }}
120
+ # {% endif %}
121
+ roles_to_train: ["assistant"]
122
+
123
+ # 4. Training Parameters
124
+ sequence_len: 2048
125
+ sample_packing: true
126
+ eval_sample_packing: true
127
+ # val_set_size: 0.05 # remove for streaming
128
+ # num_epochs: 10 # replace epochs with max_steps
129
+ max_steps: 1000 # set your target steps
130
+ dataset_prepared_path: last_run_prepared
131
+
132
+ # 5. Saving and Evaluation Strategy (use steps with streaming)
133
+ evaluation_strategy: steps
134
+ save_strategy: steps
135
+ eval_steps: 50
136
+ save_steps: 50
137
+ save_total_limit: 100
138
+
139
+ resume_from_checkpoint:
140
+
141
+ # 6. Output & Logging
142
+ output_dir: ./outputs/sft/gemma-2-2b-it-rp-sft-qlora
143
+
144
+ wandb_project: "rp-sft"
145
+ wandb_name: "gemma-2-2b-it-rp-sft-qlora"
146
+ wandb_log_model: "false"
147
+ wandb_run_id: "gemma-2-2b-it-rp-sft-qlora"
148
+
149
+ # 7. Batching & Optimizer
150
+ gradient_accumulation_steps: 4
151
+ micro_batch_size: 2
152
+ optimizer: adamw_bnb_8bit
153
+ lr_scheduler: cosine
154
+ learning_rate: 0.0002
155
+ weight_decay: 0.0
156
+
157
+ # 8. Hardware & Performance
158
+ bf16: true
159
+ #fp16: true
160
+ tf32: true
161
+
162
+ flash_attention: true
163
+ gradient_checkpointing: true
164
+ logging_steps: 1
165
+
166
+ # 9. Special Tokens
167
+ eot_tokens: ["<end_of_turn>"]
168
+ special_tokens:
169
+ bos_token: "<bos>"
170
+ eos_token: "<eos>"
171
+ pad_token: "<pad>"
172
+
173
+ ```
174
+
175
+ </details><br>
176
+
177
+ # rp-2b
178
+
179
+ This model is a fine-tuned version of [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it) on the AiAF/conversations dataset.
180
+ It achieves the following results on the evaluation set:
181
+ - Loss: 2.2455
182
+ - Memory/max Active (gib): 7.78
183
+ - Memory/max Allocated (gib): 7.78
184
+ - Memory/device Reserved (gib): 17.79
185
+
186
+ ## Model description
187
+
188
+ More information needed
189
+
190
+ ## Intended uses & limitations
191
+
192
+ More information needed
193
+
194
+ ## Training and evaluation data
195
+
196
+ More information needed
197
+
198
+ ## Training procedure
199
+
200
+ ### Training hyperparameters
201
+
202
+ The following hyperparameters were used during training:
203
+ - learning_rate: 0.0002
204
+ - train_batch_size: 1
205
+ - eval_batch_size: 2
206
+ - seed: 42
207
+ - gradient_accumulation_steps: 4
208
+ - total_train_batch_size: 4
209
+ - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
210
+ - lr_scheduler_type: cosine
211
+ - lr_scheduler_warmup_steps: 30
212
+ - training_steps: 1000
213
+
214
+ ### Training results
215
+
216
+ | Training Loss | Epoch | Step | Validation Loss | Active (gib) | Allocated (gib) | Reserved (gib) |
217
+ |:-------------:|:-----:|:----:|:---------------:|:------------:|:---------------:|:--------------:|
218
+ | No log | 0 | 0 | 3.1654 | 7.61 | 7.61 | 8.66 |
219
+ | 2.7377 | 0.05 | 50 | 2.5978 | 7.78 | 7.78 | 17.75 |
220
+ | 2.3997 | 0.1 | 100 | 2.5592 | 7.78 | 7.78 | 17.79 |
221
+ | 2.6275 | 0.15 | 150 | 2.5410 | 7.78 | 7.78 | 17.79 |
222
+ | 2.8182 | 0.2 | 200 | 2.5224 | 7.78 | 7.78 | 17.79 |
223
+ | 2.4428 | 0.25 | 250 | 2.4962 | 7.78 | 7.78 | 17.79 |
224
+ | 2.6206 | 0.3 | 300 | 2.4672 | 7.78 | 7.78 | 17.79 |
225
+ | 2.4492 | 0.35 | 350 | 2.4435 | 7.78 | 7.78 | 17.79 |
226
+ | 2.2787 | 0.4 | 400 | 2.4185 | 7.78 | 7.78 | 17.79 |
227
+ | 2.541 | 0.45 | 450 | 2.3998 | 7.78 | 7.78 | 17.79 |
228
+ | 2.5542 | 0.5 | 500 | 2.3640 | 7.78 | 7.78 | 17.79 |
229
+ | 2.6825 | 0.55 | 550 | 2.3484 | 7.78 | 7.78 | 17.79 |
230
+ | 2.6304 | 0.6 | 600 | 2.3278 | 7.78 | 7.78 | 17.79 |
231
+ | 2.4854 | 0.65 | 650 | 2.3104 | 7.78 | 7.78 | 17.79 |
232
+ | 2.3788 | 0.7 | 700 | 2.2877 | 7.78 | 7.78 | 17.79 |
233
+ | 2.2126 | 0.75 | 750 | 2.2748 | 7.78 | 7.78 | 17.79 |
234
+ | 2.4695 | 0.8 | 800 | 2.2662 | 7.78 | 7.78 | 17.79 |
235
+ | 2.5086 | 0.85 | 850 | 2.2553 | 7.78 | 7.78 | 17.79 |
236
+ | 2.404 | 0.9 | 900 | 2.2489 | 7.78 | 7.78 | 17.79 |
237
+ | 2.4012 | 0.95 | 950 | 2.2460 | 7.78 | 7.78 | 17.79 |
238
+ | 2.2586 | 1.0 | 1000 | 2.2455 | 7.78 | 7.78 | 17.79 |
239
+
240
+
241
  ### Framework versions
242
 
243
+ - PEFT 0.17.1
244
+ - Transformers 4.57.0
245
+ - Pytorch 2.7.1+cu126
246
+ - Datasets 4.0.0
247
+ - Tokenizers 0.22.1
debug.log CHANGED
@@ -5437,4 +5437,196 @@ trainable params: 83,066,880 || all params: 2,697,408,768 || trainable%: 3.0795
5437
 
5438
 
5439
  [2025-10-19 01:14:15,108] [INFO] [axolotl.core.trainers.base._save:664] [PID:42363] Saving model checkpoint to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora/checkpoint-1000
5440
 
5441
 
5442
-
 
 
 
 
 
 
 
5443
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5444
  ...adapter_model.safetensors: 18%|▏| 58.7MB / 332MB 
 
 
 
 
5445
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5446
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5447
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5448
  ...adapter_model.safetensors: 18%|▏| 58.7MB / 332MB 
 
 
 
 
5449
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5450
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5451
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5452
  ...adapter_model.safetensors: 35%|β–Ž| 117MB / 332MB 
 
 
 
 
5453
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5454
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5455
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5456
  ...adapter_model.safetensors: 50%|β–Œ| 168MB / 332MB 
 
 
 
 
5457
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5458
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5459
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5460
  ...adapter_model.safetensors: 66%|β–‹| 218MB / 332MB 
 
 
 
 
5461
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5462
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5463
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5464
  ...adapter_model.safetensors: 81%|β–Š| 268MB / 332MB 
 
 
 
 
5465
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5466
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5467
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5468
  ...adapter_model.safetensors: 96%|β–‰| 319MB / 332MB 
 
 
 
 
5469
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5470
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5471
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5472
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
 
 
 
 
5473
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5474
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5475
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5476
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
 
 
 
 
5477
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5478
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5479
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5480
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
 
 
 
 
5481
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5482
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5483
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5484
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
 
 
 
 
5485
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5486
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5487
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5488
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
 
 
 
 
5489
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5490
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5491
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
 
 
 
5492
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
 
 
 
 
5493
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
 
 
 
 
 
5494
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
 
 
5495
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB
 
5496
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB
 
5497
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB
 
5498
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB
 
5437
 
5438
 
5439
  [2025-10-19 01:14:15,108] [INFO] [axolotl.core.trainers.base._save:664] [PID:42363] Saving model checkpoint to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora/checkpoint-1000
5440
 
5441
 
5442
+
5443
+ [2025-10-19 01:14:39,753] [INFO] [axolotl.train.save_trained_model:218] [PID:42363] Training completed! Saving trained model to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora.
5444
+ [2025-10-19 01:14:40,608] [INFO] [axolotl.train.save_trained_model:336] [PID:42363] Model successfully saved to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora
5445
+ [2025-10-19 01:14:41,005] [INFO] [axolotl.core.trainers.base._save:664] [PID:42363] Saving model checkpoint to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora
5446
+
5447
+
5448
+
5449
+
5450
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5451
+
5452
+
5453
+
5454
  ...adapter_model.safetensors: 18%|▏| 58.7MB / 332MB 
5455
+
5456
+
5457
+
5458
+
5459
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5460
+
5461
+
5462
+
5463
+
5464
+
5465
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5466
+
5467
+
5468
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5469
+
5470
+
5471
+
5472
  ...adapter_model.safetensors: 18%|▏| 58.7MB / 332MB 
5473
+
5474
+
5475
+
5476
+
5477
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5478
+
5479
+
5480
+
5481
+
5482
+
5483
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5484
+
5485
+
5486
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5487
+
5488
+
5489
+
5490
  ...adapter_model.safetensors: 35%|β–Ž| 117MB / 332MB 
5491
+
5492
+
5493
+
5494
+
5495
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5496
+
5497
+
5498
+
5499
+
5500
+
5501
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5502
+
5503
+
5504
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5505
+
5506
+
5507
+
5508
  ...adapter_model.safetensors: 50%|β–Œ| 168MB / 332MB 
5509
+
5510
+
5511
+
5512
+
5513
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5514
+
5515
+
5516
+
5517
+
5518
+
5519
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5520
+
5521
+
5522
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5523
+
5524
+
5525
+
5526
  ...adapter_model.safetensors: 66%|β–‹| 218MB / 332MB 
5527
+
5528
+
5529
+
5530
+
5531
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5532
+
5533
+
5534
+
5535
+
5536
+
5537
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5538
+
5539
+
5540
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5541
+
5542
+
5543
+
5544
  ...adapter_model.safetensors: 81%|β–Š| 268MB / 332MB 
5545
+
5546
+
5547
+
5548
+
5549
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5550
+
5551
+
5552
+
5553
+
5554
+
5555
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5556
+
5557
+
5558
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5559
+
5560
+
5561
+
5562
  ...adapter_model.safetensors: 96%|β–‰| 319MB / 332MB 
5563
+
5564
+
5565
+
5566
+
5567
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5568
+
5569
+
5570
+
5571
+
5572
+
5573
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5574
+
5575
+
5576
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5577
+
5578
+
5579
+
5580
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
5581
+
5582
+
5583
+
5584
+
5585
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5586
+
5587
+
5588
+
5589
+
5590
+
5591
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5592
+
5593
+
5594
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5595
+
5596
+
5597
+
5598
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
5599
+
5600
+
5601
+
5602
+
5603
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5604
+
5605
+
5606
+
5607
+
5608
+
5609
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5610
+
5611
+
5612
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5613
+
5614
+
5615
+
5616
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
5617
+
5618
+
5619
+
5620
+
5621
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5622
+
5623
+
5624
+
5625
+
5626
+
5627
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5628
+
5629
+
5630
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5631
+
5632
+
5633
+
5634
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
5635
+
5636
+
5637
+
5638
+
5639
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5640
+
5641
+
5642
+
5643
+
5644
+
5645
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5646
+
5647
+
5648
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5649
+
5650
+
5651
+
5652
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
5653
+
5654
+
5655
+
5656
+
5657
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5658
+
5659
+
5660
+
5661
+
5662
+
5663
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5664
+
5665
+
5666
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB 
5667
+
5668
+
5669
+
5670
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB 
5671
+
5672
+
5673
+
5674
+
5675
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB 
5676
+
5677
+
5678
+
5679
+
5680
+
5681
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB 
5682
+
5683
+
5684
  ...t-qlora/training_args.bin: 100%|β–ˆ| 7.38kB / 7.38kB
5685
+
5686
  ...adapter_model.safetensors: 100%|β–ˆ| 332MB / 332MB
5687
+
5688
  ...sft-qlora/tokenizer.model: 100%|β–ˆ| 4.24MB / 4.24MB
5689
+
5690
  ...-sft-qlora/tokenizer.json: 100%|β–ˆ| 34.4MB / 34.4MB