File size: 4,877 Bytes
3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 d594870 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 d594870 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 d594870 3226a14 e7bf902 3226a14 e7bf902 3226a14 d594870 3226a14 e7bf902 d594870 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 d594870 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 e7bf902 3226a14 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 | # 🧠 ThoughtSwitch V1 1.7B Instruct — A Mode-Adaptive Reasoning Language Model
> **Model ID**: [`BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct`](https://huggingface.co/BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct)
> **Architecture**: Decoder-only transformer (GPT-style)
> **Parameters**: 1.7 Billion
> **Capabilities**: Dynamic "Thinking" vs. "Non-Thinking" mode-switching
> **Fine-Tuned for**: Instruction-following
---
## 🚀 Overview
**ThoughtSwitch V1** is a next-generation instruction-tuned language model that brings a new paradigm to text generation: **Autonomous Cognitive Mode Switching**.
It is capable of **interpreting user prompts and switching between two distinct modes** of behavior:
- 🧠 **Thinking Mode**: Deep reasoning, logical step-by-step solutions, slow but deliberate outputs.
- 💬 **Non-Thinking Mode**: Quick completions, casual replies, storytelling, and chat-like fluency.
Whether you're building reasoning agents, fast assistants, or multi-modal chains-of-thought applications, ThoughtSwitch adapts intelligently—so **you don’t have to force the prompt**.
---
## 🧠 Key Features
- ✅ **Autonomous Mode Switching**
Understands when to think deeply and when to generate fluently, based on prompt phrasing.
- ✅ **Instruction Tuned**
Trained to follow human-like instructions and align closely with user intent.
- ✅ **1.7B Parameters**
Small enough for efficient inference, yet powerful for sophisticated reasoning.
- ✅ **Open Weights**
Fully accessible under a permissive license (specify in HF model card).
---
## ✨ Example Prompts
Prompt (Thinking Mode):
"Think step by step to solve this math problem: What is 17 multiplied by 23?"
→ Reasoned output with intermediate steps and justification.
Prompt (Non-Thinking Mode):
"Write a quick sci-fi story about a robot discovering love."
→ Smooth, creative storytelling without unnecessary reasoning.
---
## 🔧 Usage
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct")
model = AutoModelForCausalLM.from_pretrained("BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct")
prompt = "Think step by step: Why does ice float on water?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=150)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
---
## 🧪 Intended Use Cases
- 🧠 **Reasoning Agents** — For multi-hop question answering, logical puzzles, or decision support.
- 📚 **Tutoring & Education** — Adaptive explanations that vary depth based on student prompts.
- 🤖 **Conversational AI** — More natural and flexible interactions with variable "thinking effort".
- ✍️ **Creative Writing** — Generate stories, poems, and ideas with or without deep context.
---
## ⚠️ Limitations
- Like all LLMs, it may hallucinate or generate biased content.
- Mode switching is **probabilistic**, not guaranteed—prompt clearly for best results.
- Performance may vary outside of English or unfamiliar domains.
---
## 📈 Performance (Unofficial Benchmarks)
| Task | Performance |
|------------------------|------------------|
| Commonsense Reasoning | ✅ Strong |
| Instruction Following | ✅ Strong |
| Fast Casual Generation | ✅ Very Strong |
| Math (Step-by-Step) | ⚠️ Moderate |
| Factual QA | ⚠️ May hallucinate|
---
## 🛠️ Model Details
- **Architecture**: GPT-style decoder (causal LM)
- **Training**: Custom pretraining with hybrid reasoning/non-reasoning dataset
- **Instruction Fine-Tuning**: Yes, using curated prompt-response pairs
- **Token Limit**: 2048 tokens (extendable with rope scaling)
---
## 🔍 Quantized Version
Looking for fast inference?
Check out the GGUF-quantized version (by @mradermacher) for compatibility with llama.cpp, KoboldAI, and other lightweight runtimes.
---
## 📄 Citation
If you use this model in your research or application, please cite it as:
@misc{thoughswitch2025,
title={ThoughtSwitch V1 1.7B Instruct: A Mode-Adaptive Reasoning Language Model},
author={BrainWave-ML},
year={2025},
howpublished={\url{https://huggingface.co/BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct}}
}
---
## 💬 Contact
For issues, feedback, or collaboration:
- 🤖 Hugging Face Page: https://huggingface.co/BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct
- 📧 Email: *[YourContact@domain.com]*
- 🌐 Website: *[https://brainwave-ml.ai]* (optional)
- 💬 Discord or Community: *Coming Soon*
---
## 🙌 Acknowledgments
Developed by the team at **BrainWave-ML**. Inspired by the question:
*“What if language models could choose when to think?”*
---
> **ThoughtSwitch**: Think when you need to. Generate when you don't.
|