End of training
Browse files
README.md
CHANGED
|
@@ -1,13 +1,247 @@
|
|
| 1 |
---
|
| 2 |
-
base_model: google/gemma-2-2b-it
|
| 3 |
library_name: peft
|
| 4 |
-
|
|
|
|
| 5 |
tags:
|
| 6 |
- axolotl
|
| 7 |
- base_model:adapter:google/gemma-2-2b-it
|
| 8 |
- lora
|
| 9 |
- transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
### Framework versions
|
| 12 |
|
| 13 |
-
- PEFT 0.17.1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
library_name: peft
|
| 3 |
+
license: gemma
|
| 4 |
+
base_model: google/gemma-2-2b-it
|
| 5 |
tags:
|
| 6 |
- axolotl
|
| 7 |
- base_model:adapter:google/gemma-2-2b-it
|
| 8 |
- lora
|
| 9 |
- transformers
|
| 10 |
+
datasets:
|
| 11 |
+
- AiAF/conversations
|
| 12 |
+
pipeline_tag: text-generation
|
| 13 |
+
model-index:
|
| 14 |
+
- name: rp-2b
|
| 15 |
+
results: []
|
| 16 |
---
|
| 17 |
+
|
| 18 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 19 |
+
should probably proofread and complete it, then remove this comment. -->
|
| 20 |
+
|
| 21 |
+
[<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
|
| 22 |
+
<details><summary>See axolotl config</summary>
|
| 23 |
+
|
| 24 |
+
axolotl version: `0.13.0.dev0`
|
| 25 |
+
```yaml
|
| 26 |
+
# 1. Base Model & Tokenizer
|
| 27 |
+
base_model: google/gemma-2-2b-it
|
| 28 |
+
model_type: AutoModelForCausalLM # Corrected from 'type_of_model' for axolotl
|
| 29 |
+
tokenizer_type: AutoTokenizer
|
| 30 |
+
hub_model_id: AiAF/rp-2b # New model ID for this finetune
|
| 31 |
+
hub_strategy: checkpoint
|
| 32 |
+
|
| 33 |
+
# 2. LoRA / QLoRA Configuration
|
| 34 |
+
load_in_4bit: true
|
| 35 |
+
adapter: qlora
|
| 36 |
+
lora_r: 64
|
| 37 |
+
lora_alpha: 128
|
| 38 |
+
lora_dropout: 0.05
|
| 39 |
+
lora_target_linear: true
|
| 40 |
+
|
| 41 |
+
# 3. Dataset Configuration (TRAIN = streamed)
|
| 42 |
+
streaming: true
|
| 43 |
+
streaming_multipack_buffer_size: 10000
|
| 44 |
+
sample_packing: true
|
| 45 |
+
datasets:
|
| 46 |
+
- path: AiAF/conversations
|
| 47 |
+
data_files: conversations_V3.jsonl
|
| 48 |
+
# revision:
|
| 49 |
+
type: chat_template
|
| 50 |
+
split: train
|
| 51 |
+
field_messages: conversations
|
| 52 |
+
message_property_mappings:
|
| 53 |
+
role: from
|
| 54 |
+
content: value
|
| 55 |
+
chat_template: jinja
|
| 56 |
+
chat_template_jinja: |
|
| 57 |
+
{{ bos_token }}
|
| 58 |
+
{% for m in messages %}
|
| 59 |
+
{% set role = 'model' if m['role']=='assistant' else 'user' %}
|
| 60 |
+
{{ '<start_of_turn>' + role + '\n' + m['content'] | trim + '<end_of_turn>\n' }}
|
| 61 |
+
{% endfor %}
|
| 62 |
+
{% if add_generation_prompt %}
|
| 63 |
+
{{ '<start_of_turn>model\n' }}
|
| 64 |
+
{% endif %}
|
| 65 |
+
|
| 66 |
+
# chat_template_jinja: |
|
| 67 |
+
# {{ bos_token }}
|
| 68 |
+
# {% set last = None %}
|
| 69 |
+
# {% for m in messages %}
|
| 70 |
+
# {% set raw_role = 'model' if m['role']=='assistant' else m['role'] %}
|
| 71 |
+
# {% set role = 'user' if raw_role=='system' else raw_role %}
|
| 72 |
+
# {% if role == last and role == 'user' %}
|
| 73 |
+
# {{ m['content'] | trim }}
|
| 74 |
+
# {% else %}
|
| 75 |
+
# {{ '<start_of_turn>' + role + '\n' + m['content'] | trim + '<end_of_turn>\n' }}
|
| 76 |
+
# {% endif %}
|
| 77 |
+
# {% set last = role %}
|
| 78 |
+
# {% endfor %}
|
| 79 |
+
# {% if add_generation_prompt %}
|
| 80 |
+
# {{ '<start_of_turn>model\n' }}
|
| 81 |
+
# {% endif %}
|
| 82 |
+
roles_to_train: ["assistant"]
|
| 83 |
+
train_on_eos: "turn"
|
| 84 |
+
# Use a fixed (non-streamed) eval file with the same schema/Jinja
|
| 85 |
+
test_datasets:
|
| 86 |
+
- path: .
|
| 87 |
+
name: json
|
| 88 |
+
type: chat_template
|
| 89 |
+
data_files: eval-datasets/shuf-1000_conversations_V2.jsonl # small, representative eval slice
|
| 90 |
+
split: train
|
| 91 |
+
field_messages: conversations
|
| 92 |
+
message_property_mappings:
|
| 93 |
+
role: from
|
| 94 |
+
content: value
|
| 95 |
+
chat_template: jinja
|
| 96 |
+
chat_template_jinja: |
|
| 97 |
+
{{ bos_token }}
|
| 98 |
+
{% for m in messages %}
|
| 99 |
+
{% set role = 'model' if m['role']=='assistant' else 'user' %}
|
| 100 |
+
{{ '<start_of_turn>' + role + '\n' + m['content'] | trim + '<end_of_turn>\n' }}
|
| 101 |
+
{% endfor %}
|
| 102 |
+
{% if add_generation_prompt %}
|
| 103 |
+
{{ '<start_of_turn>model\n' }}
|
| 104 |
+
{% endif %}
|
| 105 |
+
# chat_template_jinja: |
|
| 106 |
+
# {{ bos_token }}
|
| 107 |
+
# {% set last = None %}
|
| 108 |
+
# {% for m in messages %}
|
| 109 |
+
# {% set raw_role = 'model' if m['role']=='assistant' else m['role'] %}
|
| 110 |
+
# {% set role = 'user' if raw_role=='system' else raw_role %}
|
| 111 |
+
# {% if role == last and role == 'user' %}
|
| 112 |
+
# {{ m['content'] | trim }}
|
| 113 |
+
# {% else %}
|
| 114 |
+
# {{ '<start_of_turn>' + role + '\n' + m['content'] | trim + '<end_of_turn>\n' }}
|
| 115 |
+
# {% endif %}
|
| 116 |
+
# {% set last = role %}
|
| 117 |
+
# {% endfor %}
|
| 118 |
+
# {% if add_generation_prompt %}
|
| 119 |
+
# {{ '<start_of_turn>model\n' }}
|
| 120 |
+
# {% endif %}
|
| 121 |
+
roles_to_train: ["assistant"]
|
| 122 |
+
|
| 123 |
+
# 4. Training Parameters
|
| 124 |
+
sequence_len: 2048
|
| 125 |
+
sample_packing: true
|
| 126 |
+
eval_sample_packing: true
|
| 127 |
+
# val_set_size: 0.05 # remove for streaming
|
| 128 |
+
# num_epochs: 10 # replace epochs with max_steps
|
| 129 |
+
max_steps: 1000 # set your target steps
|
| 130 |
+
dataset_prepared_path: last_run_prepared
|
| 131 |
+
|
| 132 |
+
# 5. Saving and Evaluation Strategy (use steps with streaming)
|
| 133 |
+
evaluation_strategy: steps
|
| 134 |
+
save_strategy: steps
|
| 135 |
+
eval_steps: 50
|
| 136 |
+
save_steps: 50
|
| 137 |
+
save_total_limit: 100
|
| 138 |
+
|
| 139 |
+
resume_from_checkpoint:
|
| 140 |
+
|
| 141 |
+
# 6. Output & Logging
|
| 142 |
+
output_dir: ./outputs/sft/gemma-2-2b-it-rp-sft-qlora
|
| 143 |
+
|
| 144 |
+
wandb_project: "rp-sft"
|
| 145 |
+
wandb_name: "gemma-2-2b-it-rp-sft-qlora"
|
| 146 |
+
wandb_log_model: "false"
|
| 147 |
+
wandb_run_id: "gemma-2-2b-it-rp-sft-qlora"
|
| 148 |
+
|
| 149 |
+
# 7. Batching & Optimizer
|
| 150 |
+
gradient_accumulation_steps: 4
|
| 151 |
+
micro_batch_size: 2
|
| 152 |
+
optimizer: adamw_bnb_8bit
|
| 153 |
+
lr_scheduler: cosine
|
| 154 |
+
learning_rate: 0.0002
|
| 155 |
+
weight_decay: 0.0
|
| 156 |
+
|
| 157 |
+
# 8. Hardware & Performance
|
| 158 |
+
bf16: true
|
| 159 |
+
#fp16: true
|
| 160 |
+
tf32: true
|
| 161 |
+
|
| 162 |
+
flash_attention: true
|
| 163 |
+
gradient_checkpointing: true
|
| 164 |
+
logging_steps: 1
|
| 165 |
+
|
| 166 |
+
# 9. Special Tokens
|
| 167 |
+
eot_tokens: ["<end_of_turn>"]
|
| 168 |
+
special_tokens:
|
| 169 |
+
bos_token: "<bos>"
|
| 170 |
+
eos_token: "<eos>"
|
| 171 |
+
pad_token: "<pad>"
|
| 172 |
+
|
| 173 |
+
```
|
| 174 |
+
|
| 175 |
+
</details><br>
|
| 176 |
+
|
| 177 |
+
# rp-2b
|
| 178 |
+
|
| 179 |
+
This model is a fine-tuned version of [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it) on the AiAF/conversations dataset.
|
| 180 |
+
It achieves the following results on the evaluation set:
|
| 181 |
+
- Loss: 2.2455
|
| 182 |
+
- Memory/max Active (gib): 7.78
|
| 183 |
+
- Memory/max Allocated (gib): 7.78
|
| 184 |
+
- Memory/device Reserved (gib): 17.79
|
| 185 |
+
|
| 186 |
+
## Model description
|
| 187 |
+
|
| 188 |
+
More information needed
|
| 189 |
+
|
| 190 |
+
## Intended uses & limitations
|
| 191 |
+
|
| 192 |
+
More information needed
|
| 193 |
+
|
| 194 |
+
## Training and evaluation data
|
| 195 |
+
|
| 196 |
+
More information needed
|
| 197 |
+
|
| 198 |
+
## Training procedure
|
| 199 |
+
|
| 200 |
+
### Training hyperparameters
|
| 201 |
+
|
| 202 |
+
The following hyperparameters were used during training:
|
| 203 |
+
- learning_rate: 0.0002
|
| 204 |
+
- train_batch_size: 1
|
| 205 |
+
- eval_batch_size: 2
|
| 206 |
+
- seed: 42
|
| 207 |
+
- gradient_accumulation_steps: 4
|
| 208 |
+
- total_train_batch_size: 4
|
| 209 |
+
- optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
| 210 |
+
- lr_scheduler_type: cosine
|
| 211 |
+
- lr_scheduler_warmup_steps: 30
|
| 212 |
+
- training_steps: 1000
|
| 213 |
+
|
| 214 |
+
### Training results
|
| 215 |
+
|
| 216 |
+
| Training Loss | Epoch | Step | Validation Loss | Active (gib) | Allocated (gib) | Reserved (gib) |
|
| 217 |
+
|:-------------:|:-----:|:----:|:---------------:|:------------:|:---------------:|:--------------:|
|
| 218 |
+
| No log | 0 | 0 | 3.1654 | 7.61 | 7.61 | 8.66 |
|
| 219 |
+
| 2.7377 | 0.05 | 50 | 2.5978 | 7.78 | 7.78 | 17.75 |
|
| 220 |
+
| 2.3997 | 0.1 | 100 | 2.5592 | 7.78 | 7.78 | 17.79 |
|
| 221 |
+
| 2.6275 | 0.15 | 150 | 2.5410 | 7.78 | 7.78 | 17.79 |
|
| 222 |
+
| 2.8182 | 0.2 | 200 | 2.5224 | 7.78 | 7.78 | 17.79 |
|
| 223 |
+
| 2.4428 | 0.25 | 250 | 2.4962 | 7.78 | 7.78 | 17.79 |
|
| 224 |
+
| 2.6206 | 0.3 | 300 | 2.4672 | 7.78 | 7.78 | 17.79 |
|
| 225 |
+
| 2.4492 | 0.35 | 350 | 2.4435 | 7.78 | 7.78 | 17.79 |
|
| 226 |
+
| 2.2787 | 0.4 | 400 | 2.4185 | 7.78 | 7.78 | 17.79 |
|
| 227 |
+
| 2.541 | 0.45 | 450 | 2.3998 | 7.78 | 7.78 | 17.79 |
|
| 228 |
+
| 2.5542 | 0.5 | 500 | 2.3640 | 7.78 | 7.78 | 17.79 |
|
| 229 |
+
| 2.6825 | 0.55 | 550 | 2.3484 | 7.78 | 7.78 | 17.79 |
|
| 230 |
+
| 2.6304 | 0.6 | 600 | 2.3278 | 7.78 | 7.78 | 17.79 |
|
| 231 |
+
| 2.4854 | 0.65 | 650 | 2.3104 | 7.78 | 7.78 | 17.79 |
|
| 232 |
+
| 2.3788 | 0.7 | 700 | 2.2877 | 7.78 | 7.78 | 17.79 |
|
| 233 |
+
| 2.2126 | 0.75 | 750 | 2.2748 | 7.78 | 7.78 | 17.79 |
|
| 234 |
+
| 2.4695 | 0.8 | 800 | 2.2662 | 7.78 | 7.78 | 17.79 |
|
| 235 |
+
| 2.5086 | 0.85 | 850 | 2.2553 | 7.78 | 7.78 | 17.79 |
|
| 236 |
+
| 2.404 | 0.9 | 900 | 2.2489 | 7.78 | 7.78 | 17.79 |
|
| 237 |
+
| 2.4012 | 0.95 | 950 | 2.2460 | 7.78 | 7.78 | 17.79 |
|
| 238 |
+
| 2.2586 | 1.0 | 1000 | 2.2455 | 7.78 | 7.78 | 17.79 |
|
| 239 |
+
|
| 240 |
+
|
| 241 |
### Framework versions
|
| 242 |
|
| 243 |
+
- PEFT 0.17.1
|
| 244 |
+
- Transformers 4.57.0
|
| 245 |
+
- Pytorch 2.7.1+cu126
|
| 246 |
+
- Datasets 4.0.0
|
| 247 |
+
- Tokenizers 0.22.1
|
debug.log
CHANGED
|
@@ -5437,4 +5437,196 @@ trainable params: 83,066,880 || all params: 2,697,408,768 || trainable%: 3.0795
|
|
| 5437 |
|
| 5438 |
|
| 5439 |
[A[2025-10-19 01:14:15,108] [INFO] [axolotl.core.trainers.base._save:664] [PID:42363] Saving model checkpoint to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora/checkpoint-1000
|
| 5440 |
|
| 5441 |
|
| 5442 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5443 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5444 |
...adapter_model.safetensors: 18%|β| 58.7MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5445 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5446 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5447 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5448 |
...adapter_model.safetensors: 18%|β| 58.7MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5449 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5450 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5451 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5452 |
...adapter_model.safetensors: 35%|β| 117MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5453 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5454 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5455 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5456 |
...adapter_model.safetensors: 50%|β| 168MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5457 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5458 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5459 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5460 |
...adapter_model.safetensors: 66%|β| 218MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5461 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5462 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5463 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5464 |
...adapter_model.safetensors: 81%|β| 268MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5465 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5466 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5467 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5468 |
...adapter_model.safetensors: 96%|β| 319MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5469 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5470 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5471 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5472 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5473 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5474 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5475 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5476 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5477 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5478 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5479 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5480 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5481 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5482 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5483 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5484 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5485 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5486 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5487 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5488 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5489 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5490 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5491 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
|
|
|
|
|
|
|
|
|
| 5492 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5493 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5494 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
|
|
|
|
|
|
| 5495 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB
|
|
|
|
| 5496 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB
|
|
|
|
| 5497 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB
|
|
|
|
| 5498 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB
|
|
|
|
| 5437 |
|
| 5438 |
|
| 5439 |
[A[2025-10-19 01:14:15,108] [INFO] [axolotl.core.trainers.base._save:664] [PID:42363] Saving model checkpoint to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora/checkpoint-1000
|
| 5440 |
|
| 5441 |
|
| 5442 |
+
|
| 5443 |
+
[2025-10-19 01:14:39,753] [INFO] [axolotl.train.save_trained_model:218] [PID:42363] Training completed! Saving trained model to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora.
|
| 5444 |
+
[2025-10-19 01:14:40,608] [INFO] [axolotl.train.save_trained_model:336] [PID:42363] Model successfully saved to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora
|
| 5445 |
+
[2025-10-19 01:14:41,005] [INFO] [axolotl.core.trainers.base._save:664] [PID:42363] Saving model checkpoint to ./outputs/sft/gemma-2-2b-it-rp-sft-qlora
|
| 5446 |
+
|
| 5447 |
+
|
| 5448 |
+
|
| 5449 |
+
|
| 5450 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5451 |
+
|
| 5452 |
+
|
| 5453 |
+
|
| 5454 |
...adapter_model.safetensors: 18%|β| 58.7MB / 332MB [A[A[A
|
| 5455 |
+
|
| 5456 |
+
|
| 5457 |
+
|
| 5458 |
+
|
| 5459 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5460 |
+
|
| 5461 |
+
|
| 5462 |
+
|
| 5463 |
+
|
| 5464 |
+
|
| 5465 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5466 |
+
|
| 5467 |
+
|
| 5468 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5469 |
+
|
| 5470 |
+
|
| 5471 |
+
|
| 5472 |
...adapter_model.safetensors: 18%|β| 58.7MB / 332MB [A[A[A
|
| 5473 |
+
|
| 5474 |
+
|
| 5475 |
+
|
| 5476 |
+
|
| 5477 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5478 |
+
|
| 5479 |
+
|
| 5480 |
+
|
| 5481 |
+
|
| 5482 |
+
|
| 5483 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5484 |
+
|
| 5485 |
+
|
| 5486 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5487 |
+
|
| 5488 |
+
|
| 5489 |
+
|
| 5490 |
...adapter_model.safetensors: 35%|β| 117MB / 332MB [A[A[A
|
| 5491 |
+
|
| 5492 |
+
|
| 5493 |
+
|
| 5494 |
+
|
| 5495 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5496 |
+
|
| 5497 |
+
|
| 5498 |
+
|
| 5499 |
+
|
| 5500 |
+
|
| 5501 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5502 |
+
|
| 5503 |
+
|
| 5504 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5505 |
+
|
| 5506 |
+
|
| 5507 |
+
|
| 5508 |
...adapter_model.safetensors: 50%|β| 168MB / 332MB [A[A[A
|
| 5509 |
+
|
| 5510 |
+
|
| 5511 |
+
|
| 5512 |
+
|
| 5513 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5514 |
+
|
| 5515 |
+
|
| 5516 |
+
|
| 5517 |
+
|
| 5518 |
+
|
| 5519 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5520 |
+
|
| 5521 |
+
|
| 5522 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5523 |
+
|
| 5524 |
+
|
| 5525 |
+
|
| 5526 |
...adapter_model.safetensors: 66%|β| 218MB / 332MB [A[A[A
|
| 5527 |
+
|
| 5528 |
+
|
| 5529 |
+
|
| 5530 |
+
|
| 5531 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5532 |
+
|
| 5533 |
+
|
| 5534 |
+
|
| 5535 |
+
|
| 5536 |
+
|
| 5537 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5538 |
+
|
| 5539 |
+
|
| 5540 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5541 |
+
|
| 5542 |
+
|
| 5543 |
+
|
| 5544 |
...adapter_model.safetensors: 81%|β| 268MB / 332MB [A[A[A
|
| 5545 |
+
|
| 5546 |
+
|
| 5547 |
+
|
| 5548 |
+
|
| 5549 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5550 |
+
|
| 5551 |
+
|
| 5552 |
+
|
| 5553 |
+
|
| 5554 |
+
|
| 5555 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5556 |
+
|
| 5557 |
+
|
| 5558 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5559 |
+
|
| 5560 |
+
|
| 5561 |
+
|
| 5562 |
...adapter_model.safetensors: 96%|β| 319MB / 332MB [A[A[A
|
| 5563 |
+
|
| 5564 |
+
|
| 5565 |
+
|
| 5566 |
+
|
| 5567 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5568 |
+
|
| 5569 |
+
|
| 5570 |
+
|
| 5571 |
+
|
| 5572 |
+
|
| 5573 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5574 |
+
|
| 5575 |
+
|
| 5576 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5577 |
+
|
| 5578 |
+
|
| 5579 |
+
|
| 5580 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
| 5581 |
+
|
| 5582 |
+
|
| 5583 |
+
|
| 5584 |
+
|
| 5585 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5586 |
+
|
| 5587 |
+
|
| 5588 |
+
|
| 5589 |
+
|
| 5590 |
+
|
| 5591 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5592 |
+
|
| 5593 |
+
|
| 5594 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5595 |
+
|
| 5596 |
+
|
| 5597 |
+
|
| 5598 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
| 5599 |
+
|
| 5600 |
+
|
| 5601 |
+
|
| 5602 |
+
|
| 5603 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5604 |
+
|
| 5605 |
+
|
| 5606 |
+
|
| 5607 |
+
|
| 5608 |
+
|
| 5609 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5610 |
+
|
| 5611 |
+
|
| 5612 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5613 |
+
|
| 5614 |
+
|
| 5615 |
+
|
| 5616 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
| 5617 |
+
|
| 5618 |
+
|
| 5619 |
+
|
| 5620 |
+
|
| 5621 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5622 |
+
|
| 5623 |
+
|
| 5624 |
+
|
| 5625 |
+
|
| 5626 |
+
|
| 5627 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5628 |
+
|
| 5629 |
+
|
| 5630 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5631 |
+
|
| 5632 |
+
|
| 5633 |
+
|
| 5634 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
| 5635 |
+
|
| 5636 |
+
|
| 5637 |
+
|
| 5638 |
+
|
| 5639 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5640 |
+
|
| 5641 |
+
|
| 5642 |
+
|
| 5643 |
+
|
| 5644 |
+
|
| 5645 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5646 |
+
|
| 5647 |
+
|
| 5648 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5649 |
+
|
| 5650 |
+
|
| 5651 |
+
|
| 5652 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
| 5653 |
+
|
| 5654 |
+
|
| 5655 |
+
|
| 5656 |
+
|
| 5657 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5658 |
+
|
| 5659 |
+
|
| 5660 |
+
|
| 5661 |
+
|
| 5662 |
+
|
| 5663 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5664 |
+
|
| 5665 |
+
|
| 5666 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB [A[A
|
| 5667 |
+
|
| 5668 |
+
|
| 5669 |
+
|
| 5670 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB [A[A[A
|
| 5671 |
+
|
| 5672 |
+
|
| 5673 |
+
|
| 5674 |
+
|
| 5675 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB [A[A[A[A
|
| 5676 |
+
|
| 5677 |
+
|
| 5678 |
+
|
| 5679 |
+
|
| 5680 |
+
|
| 5681 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB [A[A[A[A[A
|
| 5682 |
+
|
| 5683 |
+
|
| 5684 |
...t-qlora/training_args.bin: 100%|β| 7.38kB / 7.38kB
|
| 5685 |
+
|
| 5686 |
...adapter_model.safetensors: 100%|β| 332MB / 332MB
|
| 5687 |
+
|
| 5688 |
...sft-qlora/tokenizer.model: 100%|β| 4.24MB / 4.24MB
|
| 5689 |
+
|
| 5690 |
...-sft-qlora/tokenizer.json: 100%|β| 34.4MB / 34.4MB
|