dragonclaw_model / train_config.json
akacaptain's picture
Upload folder using huggingface_hub
956138b verified
raw
history blame contribute delete
478 Bytes
{
"base_model": "meta-llama/Llama-3.2-3B-Instruct",
"dataset_path": "artifacts/training_data.jsonl",
"output_dir": "artifacts/model",
"lora_r": 16,
"lora_alpha": 32,
"lora_dropout": 0.05,
"backend": "hf-peft",
"dry_run": false,
"num_train_epochs": 1.0,
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 4,
"learning_rate": 0.0002,
"max_seq_length": 2048,
"warmup_ratio": 0.03,
"logging_steps": 10,
"save_steps": 200,
"seed": 42
}