training-scripts / train_qwen3_codeforces.py
kneeraj's picture
Upload train_qwen3_codeforces.py with huggingface_hub
c91de98 verified
# /// script
# dependencies = ["trl>=0.12.0", "peft>=0.7.0", "trackio", "transformers>=4.44.0", "datasets"]
# ///
from datasets import load_dataset
from peft import LoraConfig
from trl import SFTTrainer, SFTConfig
import trackio
# Load dataset with editorials for better instruction following
dataset = load_dataset(
"open-r1/codeforces-cots",
name="solutions_w_editorials_decontaminated",
split="train"
)
# Create train/eval split (90/10)
dataset_split = dataset.train_test_split(test_size=0.1, seed=42)
# LoRA configuration for efficient fine-tuning
peft_config = LoraConfig(
r=16,
lora_alpha=32,
lora_dropout=0.05,
target_modules=["q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj"],
task_type="CAUSAL_LM"
)
# SFT Training configuration
training_args = SFTConfig(
output_dir="qwen3-0.6b-codeforces-instruct",
# Training hyperparameters
num_train_epochs=3,
per_device_train_batch_size=4,
per_device_eval_batch_size=4,
gradient_accumulation_steps=4, # Effective batch size: 16
gradient_checkpointing=True,
# Learning rate and optimization
learning_rate=2e-4,
lr_scheduler_type="cosine",
warmup_ratio=0.1,
optim="paged_adamw_8bit",
# Evaluation and logging
eval_strategy="steps",
eval_steps=100,
logging_steps=10,
save_strategy="steps",
save_steps=200,
save_total_limit=3,
# Hub integration (CRITICAL - saves model to Hub)
push_to_hub=True,
hub_model_id="kneeraj/qwen3-0.6b-codeforces-instruct",
hub_strategy="every_save",
hub_private_repo=False,
# Trackio monitoring
report_to="trackio",
project="codeforces-finetuning",
run_name="qwen3-0.6b-codeforces-sft",
# Performance optimizations
bf16=True,
max_grad_norm=1.0,
# Data processing
max_seq_length=2048, # CodeForces problems can be lengthy
dataset_text_field="messages", # Use chat format
packing=False, # Don't pack for instruction following
)
# Initialize trainer
trainer = SFTTrainer(
model="Qwen/Qwen2.5-0.5B-Instruct", # Using Qwen2.5-0.5B as base (Qwen3-0.6B may not be available)
train_dataset=dataset_split["train"],
eval_dataset=dataset_split["test"],
peft_config=peft_config,
args=training_args,
)
print("Starting training...")
print(f"Training samples: {len(dataset_split['train'])}")
print(f"Evaluation samples: {len(dataset_split['test'])}")
# Train the model
trainer.train()
# Final push to Hub
print("Pushing final model to Hub...")
trainer.push_to_hub()
print("Training complete! Model saved to: kneeraj/qwen3-0.6b-codeforces-instruct")