{ "version": "v2_updated_datasets", "strategy": "smollm3_3b_lora_hard_merge", "final_loss": 0.3239577183436989, "config": { "model_path": "HuggingFaceTB/SmolLM3-3B", "output_path": "./fyodor-mini", "use_auth_token": true, "batch_size": 8, "gradient_accumulation_steps": 8, "learning_rate": 0.0002, "num_epochs": 3, "warmup_steps": 100, "max_seq_length": 1024, "mixed_precision": true, "dtype": "bfloat16" } }