File size: 4,877 Bytes
3226a14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e7bf902
 
3226a14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e7bf902
3226a14
e7bf902
d594870
e7bf902
3226a14
e7bf902
3226a14
e7bf902
3226a14
 
e7bf902
3226a14
 
 
 
e7bf902
3226a14
e7bf902
3226a14
e7bf902
3226a14
 
 
 
e7bf902
d594870
e7bf902
3226a14
e7bf902
3226a14
 
 
e7bf902
3226a14
e7bf902
3226a14
e7bf902
3226a14
 
 
 
 
 
 
e7bf902
3226a14
e7bf902
3226a14
d594870
3226a14
 
 
 
 
 
e7bf902
3226a14
e7bf902
3226a14
d594870
3226a14
e7bf902
d594870
e7bf902
3226a14
e7bf902
3226a14
e7bf902
3226a14
 
 
 
 
 
e7bf902
3226a14
 
 
 
 
e7bf902
3226a14
 
 
 
e7bf902
d594870
e7bf902
3226a14
e7bf902
3226a14
 
e7bf902
3226a14
e7bf902
3226a14
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
# 🧠 ThoughtSwitch V1 1.7B Instruct — A Mode-Adaptive Reasoning Language Model

> **Model ID**: [`BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct`](https://huggingface.co/BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct)  
> **Architecture**: Decoder-only transformer (GPT-style)  
> **Parameters**: 1.7 Billion  
> **Capabilities**: Dynamic "Thinking" vs. "Non-Thinking" mode-switching  
> **Fine-Tuned for**: Instruction-following

---

## 🚀 Overview

**ThoughtSwitch V1** is a next-generation instruction-tuned language model that brings a new paradigm to text generation: **Autonomous Cognitive Mode Switching**.

It is capable of **interpreting user prompts and switching between two distinct modes** of behavior:

- 🧠 **Thinking Mode**: Deep reasoning, logical step-by-step solutions, slow but deliberate outputs.  
- 💬 **Non-Thinking Mode**: Quick completions, casual replies, storytelling, and chat-like fluency.

Whether you're building reasoning agents, fast assistants, or multi-modal chains-of-thought applications, ThoughtSwitch adapts intelligently—so **you don’t have to force the prompt**.

---

## 🧠 Key Features

-**Autonomous Mode Switching**  
  Understands when to think deeply and when to generate fluently, based on prompt phrasing.

-**Instruction Tuned**  
  Trained to follow human-like instructions and align closely with user intent.

-**1.7B Parameters**  
  Small enough for efficient inference, yet powerful for sophisticated reasoning.

-**Open Weights**  
  Fully accessible under a permissive license (specify in HF model card).

---

## ✨ Example Prompts

Prompt (Thinking Mode):
"Think step by step to solve this math problem: What is 17 multiplied by 23?"

→ Reasoned output with intermediate steps and justification.

Prompt (Non-Thinking Mode):
"Write a quick sci-fi story about a robot discovering love."

→ Smooth, creative storytelling without unnecessary reasoning.

---

## 🔧 Usage

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct")
model = AutoModelForCausalLM.from_pretrained("BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct")

prompt = "Think step by step: Why does ice float on water?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=150)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

---

## 🧪 Intended Use Cases

- 🧠 **Reasoning Agents** — For multi-hop question answering, logical puzzles, or decision support.
- 📚 **Tutoring & Education** — Adaptive explanations that vary depth based on student prompts.
- 🤖 **Conversational AI** — More natural and flexible interactions with variable "thinking effort".
- ✍️ **Creative Writing** — Generate stories, poems, and ideas with or without deep context.

---

## ⚠️ Limitations

- Like all LLMs, it may hallucinate or generate biased content.
- Mode switching is **probabilistic**, not guaranteed—prompt clearly for best results.
- Performance may vary outside of English or unfamiliar domains.

---

## 📈 Performance (Unofficial Benchmarks)

| Task                   | Performance       |
|------------------------|------------------|
| Commonsense Reasoning  | ✅ Strong         |
| Instruction Following  | ✅ Strong         |
| Fast Casual Generation | ✅ Very Strong    |
| Math (Step-by-Step)    | ⚠️ Moderate       |
| Factual QA             | ⚠️ May hallucinate|

---

## 🛠️ Model Details

- **Architecture**: GPT-style decoder (causal LM)
- **Training**: Custom pretraining with hybrid reasoning/non-reasoning dataset
- **Instruction Fine-Tuning**: Yes, using curated prompt-response pairs
- **Token Limit**: 2048 tokens (extendable with rope scaling)

---

## 🔍 Quantized Version

Looking for fast inference?

Check out the GGUF-quantized version (by @mradermacher) for compatibility with llama.cpp, KoboldAI, and other lightweight runtimes.

---

## 📄 Citation

If you use this model in your research or application, please cite it as:

@misc{thoughswitch2025,
  title={ThoughtSwitch V1 1.7B Instruct: A Mode-Adaptive Reasoning Language Model},
  author={BrainWave-ML},
  year={2025},
  howpublished={\url{https://huggingface.co/BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct}}
}

---

## 💬 Contact

For issues, feedback, or collaboration:

- 🤖 Hugging Face Page: https://huggingface.co/BrainWave-ML/ThoughtSwitch-V1-1.7b-Instruct
- 📧 Email: *[YourContact@domain.com]*
- 🌐 Website: *[https://brainwave-ml.ai]* (optional)
- 💬 Discord or Community: *Coming Soon*

---

## 🙌 Acknowledgments

Developed by the team at **BrainWave-ML**. Inspired by the question:  
*“What if language models could choose when to think?”*

---

> **ThoughtSwitch**: Think when you need to. Generate when you don't.