--- license: apache-2.0 base_model: BoltMonkey/SuperNeuralDreadDevil-8b datasets: - Nitral-Archive/Cosmopedia-Instruct-60k-Distilled-R1-70B-ShareGPT-Reason-Tags pipeline_tag: text-generation tags: - qlora - 4bit - merged --- # boltmonkey_shortreasoning‑8b **Merged QLoRA adapter plus base weights.** Fine‑tuned for short‑form chain‑of‑thought reasoning. * **Base model:** SuperNeuralDreadDevil-8b * **Dataset:** Cosmopedia‑Instruct 60k (ShareGPT style) * **Context length:** 1096 tokens * **Training:** 4 epochs, LoRA r = 32, α = 16, dropout 0.05, fp16, 4‑bit quant See `train_args.json` for the full Axolotl config.