| model: | |
| base_model: deepseek-ai/deepseek-math-7b-instruct | |
| trust_remote_code: true | |
| use_bf16: true | |
| load_in_4bit: true | |
| bnb_4bit_quant_type: nf4 | |
| bnb_4bit_use_double_quant: true | |
| attn_implementation: null | |
| lora: | |
| r: 64 | |
| alpha: 128 | |
| dropout: 0.05 | |
| bias: none | |
| target_modules: | |
| - q_proj | |
| - k_proj | |
| - v_proj | |
| - o_proj | |
| - gate_proj | |
| - up_proj | |
| - down_proj | |
| data: | |
| train_file: data/releases/v1/train.parquet | |
| validation_file: data/releases/v1/validation.parquet | |
| prompt_field: prompt | |
| target_field: target | |
| final_answer_field: final_answer | |
| proof_field: proof_formal | |
| max_seq_length: 2048 | |
| max_train_samples: null | |
| max_eval_samples: 2000 | |
| system_prompt: | | |
| You are a rigorous mathematical reasoning assistant specialized in unsolved | |
| conjectures. Produce clear, checkable reasoning and avoid claiming a full | |
| proof unless it is explicitly available in the task context. | |
| training: | |
| output_dir: model_development/runs/deepseek-math-lora | |
| num_train_epochs: 1 | |
| per_device_train_batch_size: 1 | |
| per_device_eval_batch_size: 1 | |
| gradient_accumulation_steps: 16 | |
| learning_rate: 2.0e-5 | |
| weight_decay: 0.01 | |
| warmup_ratio: 0.03 | |
| lr_scheduler_type: cosine | |
| max_grad_norm: 1.0 | |
| gradient_checkpointing: true | |
| logging_steps: 10 | |
| save_steps: 250 | |
| eval_steps: 250 | |
| save_total_limit: 3 | |
| dataloader_num_workers: 2 | |
| seed: 17 | |
| hub: | |
| push_to_hub: true | |
| repo_id: NorthernTribe-Research/math-conjecture-model | |
| private: false | |
| commit_message: Train DeepSeek-Math LoRA on conjecture corpus. | |
| credentials: | |
| path: huggingface-api-key.json | |