ArunkumarVR commited on
Commit
6ac79a5
·
verified ·
1 Parent(s): aee4ff6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -60,7 +60,7 @@ We publish **supported releases, experimental variants, and research checkpoints
60
 
61
  **DeepBrainz-R1-4B** is a compact, high-performance reasoning model engineered by **DeepBrainz AI & Labs**. It is part of the **DeepBrainz-R1 Series**, designed to deliver frontier-class reasoning capabilities in cost-effective parameter sizes.
62
 
63
- This specific variant offers a **32,768 token context window**, making it suitable for `extended context version optimized for medium-length document analysis.`.
64
 
65
  ---
66
 
@@ -82,7 +82,8 @@ This specific variant offers a **32,768 token context window**, making it suitab
82
  - **Code Generation:** Writing and debugging algorithms.
83
  - **Structured Data Extraction:** Parsing and reasoning over unstructured text.
84
 
85
- > **Note:** This is a base reasoning model. For conversational chat, we recommend using a specific instruct template or fine-tuning on your domain data.
 
86
 
87
  ---
88
 
@@ -111,9 +112,9 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
111
 
112
  ## 🏗️ Technical Summary
113
 
114
- The model was produced using a **multi-stage optimization process** involving large-scale supervision and iterative refinement. It is designed to maximize reasoning quality while maintaining instruction robustness.
115
 
116
- *Specific training methodologies and dataset compositions are proprietary.*
117
 
118
  ---
119
 
 
60
 
61
  **DeepBrainz-R1-4B** is a compact, high-performance reasoning model engineered by **DeepBrainz AI & Labs**. It is part of the **DeepBrainz-R1 Series**, designed to deliver frontier-class reasoning capabilities in cost-effective parameter sizes.
62
 
63
+ This variant offers an extended context window (up to 32,768 tokens), making it suitable for medium-length document and code analysis.
64
 
65
  ---
66
 
 
82
  - **Code Generation:** Writing and debugging algorithms.
83
  - **Structured Data Extraction:** Parsing and reasoning over unstructured text.
84
 
85
+ > **Note:** This model has undergone post-training to enhance reasoning quality and agentic reliability.
86
+ > It is not optimized for open-ended conversational chat without additional instruction tuning.
87
 
88
  ---
89
 
 
112
 
113
  ## 🏗️ Technical Summary
114
 
115
+ The model has undergone **post-training** to improve reasoning quality, output stability, and robustness under agentic workloads.
116
 
117
+ *Detailed post-training recipes and dataset compositions are not fully disclosed.*
118
 
119
  ---
120