gr0010 commited on
Commit
57c90d1
·
verified ·
1 Parent(s): c2ebe45

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -1,11 +1,11 @@
1
  ---
2
  license: apache-2.0
3
  ---
4
- # Art-0-8B: _Reasoning the way you want it to with Adaptive Thinking_
5
 
6
- **[Join Our Community](https://agi-0.com)**
7
 
8
- Art-0-8B is the first open-source LLM that allows users to explicitly control its reasoning methodology through direct prompting instructions.
9
 
10
  This experimental model is fine-tuned on Qwen3-8B using a specialized dataset that makes the model's thinking style directly controllable through system prompts, similar to how you would instruct an LLM to adopt a specific persona or output format.
11
 
@@ -21,20 +21,20 @@ Reasoning Instructions:
21
 
22
  While the model is primarily trained to implement adaptive thinking based on system prompt instructions, it can also respond to reasoning style changes requested during mid-conversation, though this functionality may not be consistently reliable.
23
 
24
- Some of the benefits that Adaptive Thinking enables:
25
 
26
  - Direct control over AI reasoning patterns and output structure
27
  - Enhanced experimentation with reasoning models and potential for RL strategies that optimize thinking styles
28
  - Improved safety through explicit control over the reasoning process
29
  - Customizable thinking approaches tailored to specific tasks
30
 
31
- While prompt engineering has long been known to improve LLM performance, Art-0 represents the first approach that gives users direct control over the internal reasoning process of LLMs. This capability allows models to be configured with optimal thinking patterns for different use cases.
32
 
33
  _If you like this, please consider leaving a like on the repository—it would help us, and if you can, also leave feedback in the community section._
34
 
35
  ## 🎯 See Art in Action
36
 
37
- [**→ Try more examples in our interactive demo**](https://huggingface.co/spaces/gr0010/Try-Art-0-8B)
38
  ### Example: Thinking in Rap Lyrics
39
 
40
  **System Prompt:**
@@ -123,14 +123,14 @@ The number of R's in "raspberry" is 3. The fourth prime number is 7. Multiplying
123
  **Answer: 441**
124
  ```
125
 
126
- [**→ Try more examples in our interactive demo**](https://huggingface.co/spaces/gr0010/Try-Art-0-8B)
127
 
128
  ## 🚀 Quick Start
129
 
130
  ```python
131
  from transformers import AutoModelForCausalLM, AutoTokenizer
132
 
133
- model_name = "AGI-0/Art-0-8B"
134
 
135
  # load the tokenizer and the model
136
  tokenizer = AutoTokenizer.from_pretrained(model_name)
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+ # CustomThinker-0-8B: _Reasoning the way you want it to_
5
 
6
+ > If you want to contact me about this experiment, consulting, or anything else you can find my email at https://gr.bio
7
 
8
+ CustomThinker-0-8B is the first open-source LLM that allows users to explicitly control its reasoning methodology through direct prompting instructions.
9
 
10
  This experimental model is fine-tuned on Qwen3-8B using a specialized dataset that makes the model's thinking style directly controllable through system prompts, similar to how you would instruct an LLM to adopt a specific persona or output format.
11
 
 
21
 
22
  While the model is primarily trained to implement adaptive thinking based on system prompt instructions, it can also respond to reasoning style changes requested during mid-conversation, though this functionality may not be consistently reliable.
23
 
24
+ Some of the benefits that this strategy enables:
25
 
26
  - Direct control over AI reasoning patterns and output structure
27
  - Enhanced experimentation with reasoning models and potential for RL strategies that optimize thinking styles
28
  - Improved safety through explicit control over the reasoning process
29
  - Customizable thinking approaches tailored to specific tasks
30
 
31
+ While prompt engineering has long been known to improve LLM performance, CustomThinker represents the first approach that gives users direct control over the internal reasoning process of LLMs. This capability allows models to be configured with optimal thinking patterns for different use cases.
32
 
33
  _If you like this, please consider leaving a like on the repository—it would help us, and if you can, also leave feedback in the community section._
34
 
35
  ## 🎯 See Art in Action
36
 
37
+ [**→ Try more examples in our interactive demo**](https://huggingface.co/spaces/gr0010/CustomThinker-Demo)
38
  ### Example: Thinking in Rap Lyrics
39
 
40
  **System Prompt:**
 
123
  **Answer: 441**
124
  ```
125
 
126
+ [**→ Try more examples in the interactive demo**](https://huggingface.co/spaces/gr0010/Try-Art-0-8B)
127
 
128
  ## 🚀 Quick Start
129
 
130
  ```python
131
  from transformers import AutoModelForCausalLM, AutoTokenizer
132
 
133
+ model_name = "gr0010/CustomThinker-0-8B"
134
 
135
  # load the tokenizer and the model
136
  tokenizer = AutoTokenizer.from_pretrained(model_name)