--- base_model: Qwen/Qwen3-30B-A3B library_name: peft pipeline_tag: text-generation tags: - prefix-tuning - persona - einstein - philosophy - debate --- # Einstein Prefix Adapter Prefix-tuned adapter that teaches the model to embody Albert Einstein's reasoning patterns, voice, and philosophical positions. ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3-30B-A3B") model = PeftModel.from_pretrained(base_model, "debaterhub/prefix-einstein") tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen3-30B-A3B") ``` ## Training Details - **Method**: PEFT Prefix-Tuning (40 virtual tokens) - **Base Model**: Qwen/Qwen3-30B-A3B (30B MoE, 3B active) - **Dataset**: 447 examples of Einstein-style debate responses - **Epochs**: 3 - **Hardware**: 8x A100-40GB ## Evaluation Evaluated using LLM-as-Judge (Claude Opus 4.5) on 5 dimensions: - Ideational Fidelity (35%) - Reasoning Pattern (25%) - Voice Authenticity (20%) - Engagement Quality (15%) - Anti-Patterns (5%) **Baseline**: 3.3/5.0 **Trained**: 3.4/5.0 Key improvement: Reduced meta-roleplay anti-patterns, more direct in-character responses. ## Framework Versions - PEFT 0.18.0 - Transformers 4.46.0