[2025-10-22 12:55:24,619] [DEBUG] [axolotl.utils.config.resolve_dtype:66] [PID:2418] bf16 support detected, enabling for this configuration. config.json: 0%| | 0.00/663 [00:00 [2025-10-22 12:55:25,792] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:279] [PID:2418] BOS: None / None [2025-10-22 12:55:25,792] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:280] [PID:2418] PAD: 151643 / <|endoftext|> [2025-10-22 12:55:25,792] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:281] [PID:2418] UNK: None / None [2025-10-22 12:55:25,792] [INFO] [axolotl.utils.data.shared.load_preprocessed_dataset:476] [PID:2418] Unable to find prepared dataset in last_run_prepared/a99a12059c50ab085817560a37dbde6c [2025-10-22 12:55:25,792] [INFO] [axolotl.utils.data.sft._load_raw_datasets:320] [PID:2418] Loading raw datasets... [2025-10-22 12:55:25,792] [WARNING] [axolotl.utils.data.sft._load_raw_datasets:322] [PID:2418] Processing datasets during training can lead to VRAM instability. Please pre-process your dataset using `axolotl preprocess path/to/config.yml`. Generating train split: 0 examples [00:00, ? examples/s] Generating train split: 333 examples [00:00, 21071.50 examples/s] [2025-10-22 12:55:25,923] [INFO] [axolotl.utils.data.wrappers.get_dataset_wrapper:87] [PID:2418] Loading dataset: /workspace/fine-tuning/data/data.json with base_type: alpaca and prompt_style: None Tokenizing Prompts (num_proc=36): 0%| | 0/333 [00:004096) (num_proc=36): 0%| | 0/333 [00:004096) (num_proc=36): 3%|█▊ | 10/333 [00:00<00:27, 11.56 examples/s] Dropping Long Sequences (>4096) (num_proc=36): 75%|████████████████████████████████████████████▍ | 251/333 [00:00<00:00, 345.84 examples/s] Dropping Long Sequences (>4096) (num_proc=36): 100%|███████████████████████████████████████████████████████████| 333/333 [00:01<00:00, 273.41 examples/s] Saving the dataset (0/1 shards): 0%| | 0/333 [00:00 [2025-10-22 12:55:32,147] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:279] [PID:2418] BOS: None / None [2025-10-22 12:55:32,147] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:280] [PID:2418] PAD: 151643 / <|endoftext|> [2025-10-22 12:55:32,147] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:281] [PID:2418] UNK: None / None [2025-10-22 12:55:32,147] [DEBUG] [axolotl.train.setup_model_and_tokenizer:74] [PID:2418] Loading model [2025-10-22 12:55:32,205] [DEBUG] [axolotl.monkeypatch.transformers.trainer_loss_calc.patch_evaluation_loop:87] [PID:2418] Patched Trainer.evaluation_loop with nanmean loss calculation [2025-10-22 12:55:32,206] [DEBUG] [axolotl.monkeypatch.transformers.trainer_loss_calc.patch_maybe_log_save_evaluate:138] [PID:2418] Patched Trainer._maybe_log_save_evaluate with nanmean loss calculation model.safetensors.index.json: 0.00B [00:00, ?B/s] model.safetensors.index.json: 27.8kB [00:00, 193MB/s] model-00001-of-00004.safetensors: 0%| | 0.00/3.95G [00:00