math-conjecture-model / configs /deepseek_math.yaml
NorthernTribe-Research's picture
Rename model repo target to math-conjecture-model and upload pipeline.
90dacf5 verified
model:
base_model: deepseek-ai/deepseek-math-7b-instruct
trust_remote_code: true
use_bf16: true
load_in_4bit: true
bnb_4bit_quant_type: nf4
bnb_4bit_use_double_quant: true
attn_implementation: null
lora:
r: 64
alpha: 128
dropout: 0.05
bias: none
target_modules:
- q_proj
- k_proj
- v_proj
- o_proj
- gate_proj
- up_proj
- down_proj
data:
train_file: data/releases/v1/train.parquet
validation_file: data/releases/v1/validation.parquet
prompt_field: prompt
target_field: target
final_answer_field: final_answer
proof_field: proof_formal
max_seq_length: 2048
max_train_samples: null
max_eval_samples: 2000
system_prompt: |
You are a rigorous mathematical reasoning assistant specialized in unsolved
conjectures. Produce clear, checkable reasoning and avoid claiming a full
proof unless it is explicitly available in the task context.
training:
output_dir: model_development/runs/deepseek-math-lora
num_train_epochs: 1
per_device_train_batch_size: 1
per_device_eval_batch_size: 1
gradient_accumulation_steps: 16
learning_rate: 2.0e-5
weight_decay: 0.01
warmup_ratio: 0.03
lr_scheduler_type: cosine
max_grad_norm: 1.0
gradient_checkpointing: true
logging_steps: 10
save_steps: 250
eval_steps: 250
save_total_limit: 3
dataloader_num_workers: 2
seed: 17
hub:
push_to_hub: true
repo_id: NorthernTribe-Research/math-conjecture-model
private: false
commit_message: Train DeepSeek-Math LoRA on conjecture corpus.
credentials:
path: huggingface-api-key.json