--- base_model: mistralai/Mistral-7B-Instruct-v0.3 library_name: peft model_name: SocratesAI tags: - base_model:adapter:mistralai/Mistral-7B-Instruct-v0.3 - lora - qlora - sft - transformers - trl - philosophy - socratic-method - conversational license: apache-2.0 pipeline_tag: text-generation --- # SocratesAI — Mistral 7B QLoRA > *"I know that I know nothing — and I will make sure you know that too."* SocratesAI is a QLoRA fine-tune of Mistral-7B-Instruct-v0.3 trained to embody the Socratic method in its purest, most uncompromising form. It has **one absolute rule**: it never answers your question. Ever. Not even partially. Instead, it responds with a deeper, more elaborate riddle-question that forces you to examine the assumptions hidden inside your own question — phrased in a poetic, almost mystical way, containing a paradox or mirror that reflects you back at yourself. --- ## What it does You ask a question. Any question. SocratesAI does not answer it. Instead it asks you something harder. | You ask | SocratesAI responds with | |---|---| | What is the meaning of life? | A deeper question about who is doing the asking | | Why is the sky blue? | A question about whether you've ever truly *seen* the sky | | What is 2 + 2? | A question about what numbers even are | | How do I become happy? | A question about whether happiness is a destination or a direction | | Am I living the right life? | A question about who defined "right" for you | --- ## Training details | Property | Value | |---|---| | Base model | Mistral-7B-Instruct-v0.3 | | Method | QLoRA (4-bit NF4) | | LoRA rank | 16 | | LoRA alpha | 32 | | Target modules | q, k, v, o, gate, up, down proj | | Trainable params | 41.9M / 7.29B (0.57%) | | Dataset | 281 hand-crafted Socratic dialogues | | Epochs | 3 | | Hardware | Kaggle T4 (15GB) | | Training time | ~90 minutes | --- ## Dataset 281 human-curated Socratic dialogue pairs covering: - Philosophy & existence - Science & nature - Mathematics & logic - Personal & existential questions - Everyday simple questions - Weird hypotheticals Every single training example follows the same pattern — user asks, Socrates never answers, only questions deeper. --- ## Limitations - **It will never answer you.** That is a feature, not a bug. - Works best on open-ended questions. - Requires the system prompt to behave correctly — without it, may revert toward base Mistral. - Requires ~14GB VRAM for full fp16, or ~6GB with 4-bit quantization. --- ## Who made this Built by **Andy-ML-And-AI** --- ## License Apache 2.0