CRD_V7 - Clinical Reasoning Model (Research-Optimized)
Overview
Fine-tuned Llama 3.1 8B on Clinical Reasoning Dataset using research-optimized hyperparameters based on Med42 clinical LLM study and "LoRA Without Regret" findings.
Dataset
- Source: Clinical Reasoning Data (crd_medprompt_unsloth.json)
- Examples: 10,081
- Format: Structured reasoning with "Key facts โ Mechanism โ Why-not alternatives โ "
- Avg Tokens: ~700 per example
Research-Optimized Configuration
| Parameter | Value | Research Basis |
|---|---|---|
| Learning Rate | 0.0001 | Med42: 1e-4 optimal for clinical |
| LoRA Rank | 64 | Higher rank for complex patterns |
| LoRA Alpha | 128 | 2ร rank standard |
| LoRA Dropout | 0.05 | Regularization for structured output |
| Target Modules | All (MLP + Attention) | MLP critical for reasoning |
| Epochs | 3 | Structured formats need more exposure |
| Batch Size | 16 | Research-optimal for LoRA |
| Warmup | 5% | Stability for complex learning |
| RSLoRA | Enabled | Rank-stabilized for rโฅ64 |
Training Results
- Training Time: 2.66 hours
- Final Loss: 0.5773
Part of Experiment
Testing hypothesis: Clinical reasoning data improves general reasoning capabilities.
- kinzakhan1/CRD_V7 - 100% Clinical (this model)
- kinzakhan1/SRD_V7 - 100% Standard reasoning
- kinzakhan1/MIXED_V7 - 50/50 mix
- Downloads last month
- 44
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for kinzakhan1/CRD_V7
Base model
meta-llama/Llama-3.1-8B
Finetuned
meta-llama/Llama-3.1-8B-Instruct