Update README.md
Browse files
README.md
CHANGED
|
@@ -19,8 +19,8 @@ tags:
|
|
| 19 |
|
| 20 |
# **OpenScienceReasoning-Qwen-e10**
|
| 21 |
|
| 22 |
-
>
|
| 23 |
-
>
|
| 24 |
|
| 25 |
> \[!note]
|
| 26 |
> GGUF: [https://huggingface.co/prithivMLmods/OpenScienceReasoning-Qwen-e10-GGUF](https://huggingface.co/prithivMLmods/OpenScienceReasoning-Qwen-e10-GGUF)
|
|
@@ -29,23 +29,23 @@ tags:
|
|
| 29 |
|
| 30 |
## **Key Features**
|
| 31 |
|
| 32 |
-
1. **
|
| 33 |
-
Fine-tuned on **
|
| 34 |
|
| 35 |
-
2. **
|
| 36 |
-
|
| 37 |
|
| 38 |
-
3. **
|
| 39 |
-
Performs
|
| 40 |
|
| 41 |
-
4. **
|
| 42 |
-
|
| 43 |
|
| 44 |
-
5. **
|
| 45 |
-
|
| 46 |
|
| 47 |
-
6. **Optimized for
|
| 48 |
-
|
| 49 |
|
| 50 |
---
|
| 51 |
|
|
@@ -63,10 +63,10 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
| 63 |
)
|
| 64 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 65 |
|
| 66 |
-
prompt = "Explain
|
| 67 |
|
| 68 |
messages = [
|
| 69 |
-
{"role": "system", "content": "You are a scientific tutor skilled in reasoning, math, and
|
| 70 |
{"role": "user", "content": prompt}
|
| 71 |
]
|
| 72 |
|
|
@@ -94,15 +94,15 @@ print(response)
|
|
| 94 |
|
| 95 |
## **Intended Use**
|
| 96 |
|
| 97 |
-
* Scientific
|
| 98 |
-
*
|
| 99 |
-
* Structured technical
|
| 100 |
-
*
|
| 101 |
-
* Deployment in
|
| 102 |
|
| 103 |
## **Limitations**
|
| 104 |
|
| 105 |
-
* Not
|
| 106 |
-
*
|
| 107 |
-
*
|
| 108 |
-
*
|
|
|
|
| 19 |
|
| 20 |
# **OpenScienceReasoning-Qwen-e10**
|
| 21 |
|
| 22 |
+
> OpenScienceReasoning-Qwen-e10 is a high-efficiency, science-focused reasoning model fine-tuned on **Qwen3-1.7B** using the [**nvidia/OpenScienceReasoning-2**](https://huggingface.co/datasets/nvidia/OpenScienceReasoning-2) dataset. It incorporates **10,000 distinct entries** for scientific reasoning, chain-of-thought exploration, and analytical problem solving.
|
| 23 |
+
> The model blends symbolic precision, scientific logic, and structured output fluency—making it an ideal tool for researchers, educators, and developers seeking advanced reasoning under constrained compute.
|
| 24 |
|
| 25 |
> \[!note]
|
| 26 |
> GGUF: [https://huggingface.co/prithivMLmods/OpenScienceReasoning-Qwen-e10-GGUF](https://huggingface.co/prithivMLmods/OpenScienceReasoning-Qwen-e10-GGUF)
|
|
|
|
| 29 |
|
| 30 |
## **Key Features**
|
| 31 |
|
| 32 |
+
1. **Scientific Reasoning & Chain-of-Thought**
|
| 33 |
+
Fine-tuned on **10,000 curated entries** from the **OpenScienceReasoning-2** dataset, designed to enhance step-by-step analytical reasoning in science and mathematics.
|
| 34 |
|
| 35 |
+
2. **Advanced Code Reasoning & Generation**
|
| 36 |
+
Supports multi-language coding with explanations, optimization hints, and error detection—ideal for algorithm synthesis, debugging, and prototyping.
|
| 37 |
|
| 38 |
+
3. **Mathematical & Scientific Problem Solving**
|
| 39 |
+
Performs analytical reasoning in physics, biology, chemistry, and mathematics—explaining concepts, solving equations, and handling symbolic derivations.
|
| 40 |
|
| 41 |
+
4. **Hybrid Symbolic-AI Thinking**
|
| 42 |
+
Combines structured logic, chain-of-thought reasoning, and open-ended inference, delivering robust performance on STEM-related tasks.
|
| 43 |
|
| 44 |
+
5. **Structured Output Mastery**
|
| 45 |
+
Seamlessly generates output in **LaTeX**, **Markdown**, **JSON**, **CSV**, and **YAML**, suited for technical documentation, research papers, and structured data.
|
| 46 |
|
| 47 |
+
6. **Optimized Lightweight Footprint for Versatile Deployment**
|
| 48 |
+
Balances performance and efficiency, making it deployable on **mid-range GPUs**, **offline clusters**, and **edge AI systems**.
|
| 49 |
|
| 50 |
---
|
| 51 |
|
|
|
|
| 63 |
)
|
| 64 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 65 |
|
| 66 |
+
prompt = "Explain the difference between Newtonian mechanics and quantum mechanics with examples."
|
| 67 |
|
| 68 |
messages = [
|
| 69 |
+
{"role": "system", "content": "You are a scientific tutor skilled in reasoning, math, and coding."},
|
| 70 |
{"role": "user", "content": prompt}
|
| 71 |
]
|
| 72 |
|
|
|
|
| 94 |
|
| 95 |
## **Intended Use**
|
| 96 |
|
| 97 |
+
* Scientific tutoring, computational reasoning, and mathematical education
|
| 98 |
+
* Research assistant for physics, chemistry, biology, and interdisciplinary domains
|
| 99 |
+
* Structured technical data generation in multiple formats
|
| 100 |
+
* STEM-focused chatbot or API for research and education tools
|
| 101 |
+
* Deployment in mid-resource environments requiring high reasoning fidelity
|
| 102 |
|
| 103 |
## **Limitations**
|
| 104 |
|
| 105 |
+
* Not tuned for general-purpose or long-form creative writing
|
| 106 |
+
* Context limitations may hinder multi-document or full codebase analysis
|
| 107 |
+
* Specialized for scientific and technical reasoning—general chat may underperform
|
| 108 |
+
* Prioritizes structured logic over casual or emotional tone generation
|