Update README.md
Browse files
README.md
CHANGED
|
@@ -1,11 +1,11 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
---
|
| 4 |
-
#
|
| 5 |
|
| 6 |
-
|
| 7 |
|
| 8 |
-
|
| 9 |
|
| 10 |
This experimental model is fine-tuned on Qwen3-8B using a specialized dataset that makes the model's thinking style directly controllable through system prompts, similar to how you would instruct an LLM to adopt a specific persona or output format.
|
| 11 |
|
|
@@ -21,20 +21,20 @@ Reasoning Instructions:
|
|
| 21 |
|
| 22 |
While the model is primarily trained to implement adaptive thinking based on system prompt instructions, it can also respond to reasoning style changes requested during mid-conversation, though this functionality may not be consistently reliable.
|
| 23 |
|
| 24 |
-
Some of the benefits that
|
| 25 |
|
| 26 |
- Direct control over AI reasoning patterns and output structure
|
| 27 |
- Enhanced experimentation with reasoning models and potential for RL strategies that optimize thinking styles
|
| 28 |
- Improved safety through explicit control over the reasoning process
|
| 29 |
- Customizable thinking approaches tailored to specific tasks
|
| 30 |
|
| 31 |
-
While prompt engineering has long been known to improve LLM performance,
|
| 32 |
|
| 33 |
_If you like this, please consider leaving a like on the repository—it would help us, and if you can, also leave feedback in the community section._
|
| 34 |
|
| 35 |
## 🎯 See Art in Action
|
| 36 |
|
| 37 |
-
[**→ Try more examples in our interactive demo**](https://huggingface.co/spaces/gr0010/
|
| 38 |
### Example: Thinking in Rap Lyrics
|
| 39 |
|
| 40 |
**System Prompt:**
|
|
@@ -123,14 +123,14 @@ The number of R's in "raspberry" is 3. The fourth prime number is 7. Multiplying
|
|
| 123 |
**Answer: 441**
|
| 124 |
```
|
| 125 |
|
| 126 |
-
[**→ Try more examples in
|
| 127 |
|
| 128 |
## 🚀 Quick Start
|
| 129 |
|
| 130 |
```python
|
| 131 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 132 |
|
| 133 |
-
model_name = "
|
| 134 |
|
| 135 |
# load the tokenizer and the model
|
| 136 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
---
|
| 4 |
+
# CustomThinker-0-8B: _Reasoning the way you want it to_
|
| 5 |
|
| 6 |
+
> If you want to contact me about this experiment, consulting, or anything else you can find my email at https://gr.bio
|
| 7 |
|
| 8 |
+
CustomThinker-0-8B is the first open-source LLM that allows users to explicitly control its reasoning methodology through direct prompting instructions.
|
| 9 |
|
| 10 |
This experimental model is fine-tuned on Qwen3-8B using a specialized dataset that makes the model's thinking style directly controllable through system prompts, similar to how you would instruct an LLM to adopt a specific persona or output format.
|
| 11 |
|
|
|
|
| 21 |
|
| 22 |
While the model is primarily trained to implement adaptive thinking based on system prompt instructions, it can also respond to reasoning style changes requested during mid-conversation, though this functionality may not be consistently reliable.
|
| 23 |
|
| 24 |
+
Some of the benefits that this strategy enables:
|
| 25 |
|
| 26 |
- Direct control over AI reasoning patterns and output structure
|
| 27 |
- Enhanced experimentation with reasoning models and potential for RL strategies that optimize thinking styles
|
| 28 |
- Improved safety through explicit control over the reasoning process
|
| 29 |
- Customizable thinking approaches tailored to specific tasks
|
| 30 |
|
| 31 |
+
While prompt engineering has long been known to improve LLM performance, CustomThinker represents the first approach that gives users direct control over the internal reasoning process of LLMs. This capability allows models to be configured with optimal thinking patterns for different use cases.
|
| 32 |
|
| 33 |
_If you like this, please consider leaving a like on the repository—it would help us, and if you can, also leave feedback in the community section._
|
| 34 |
|
| 35 |
## 🎯 See Art in Action
|
| 36 |
|
| 37 |
+
[**→ Try more examples in our interactive demo**](https://huggingface.co/spaces/gr0010/CustomThinker-Demo)
|
| 38 |
### Example: Thinking in Rap Lyrics
|
| 39 |
|
| 40 |
**System Prompt:**
|
|
|
|
| 123 |
**Answer: 441**
|
| 124 |
```
|
| 125 |
|
| 126 |
+
[**→ Try more examples in the interactive demo**](https://huggingface.co/spaces/gr0010/Try-Art-0-8B)
|
| 127 |
|
| 128 |
## 🚀 Quick Start
|
| 129 |
|
| 130 |
```python
|
| 131 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 132 |
|
| 133 |
+
model_name = "gr0010/CustomThinker-0-8B"
|
| 134 |
|
| 135 |
# load the tokenizer and the model
|
| 136 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|