Update README.md
Browse files
README.md
CHANGED
|
@@ -10,17 +10,11 @@ datasets:
|
|
| 10 |
|
| 11 |
<img src='cheap.png' width='700'>
|
| 12 |
|
| 13 |
-
|
| 14 |
-
**A 4B-parameter Qwen model distilled from Tongyi DeepResearch-30B A3B**, optimized for web-scale “deep research” tasks and plug-and-play inference with **[Alibaba-NLP/DeepResearch](https://github.com/Alibaba-NLP/DeepResearch)**.
|
| 15 |
-
|
| 16 |
[](https://huggingface.co/your-username/your-model-name)
|
| 17 |
[](#license)
|
| 18 |
[](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
|
| 19 |
|
| 20 |
-
|
| 21 |
-
---
|
| 22 |
-
|
| 23 |
-
## TL;DR
|
| 24 |
|
| 25 |
* **Base**: Qwen 4B (dense)
|
| 26 |
* **Teacher**: Tongyi DeepResearch 30B A3B (MoE)
|
|
@@ -40,7 +34,6 @@ datasets:
|
|
| 40 |
|
| 41 |
* **Primary dataset**: [`cheapresearch/CheapResearch-DS-33k`](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
|
| 42 |
|
| 43 |
-
|
| 44 |
---
|
| 45 |
|
| 46 |
## Inference with Alibaba-NLP/DeepResearch (Recommended)
|
|
@@ -65,22 +58,16 @@ Edit the config to add this model
|
|
| 65 |
MODEL_PATH=cheapresearch/CheapResearch-4B-Thinking
|
| 66 |
```
|
| 67 |
|
| 68 |
-
> ⚠️ **Note**: Use a **search-enabled** profile in DeepResearch so the model can browse and cite sources. Disable “reasoning suppression” features—this student is trained to produce compact but explicit research traces.
|
| 69 |
-
|
| 70 |
### Hardware notes
|
| 71 |
|
| 72 |
-
* **Single
|
| 73 |
|
| 74 |
---
|
| 75 |
|
| 76 |
## Evaluation
|
| 77 |
|
| 78 |
-
|
| 79 |
-
|
| 80 |
-
| HLE textonly 200 @1 | Correctness (o4) | — | — | With HLE keyword filtering to prevent cheating |
|
| 81 |
-
| SimpleQA @1 | Win-Rate vs Baseline | Correctness (o4) | — | With SimpleQA keyword filtering to prevent cheating |
|
| 82 |
-
|
| 83 |
-
|
| 84 |
|
| 85 |
## Acknowledgements
|
| 86 |
|
|
|
|
| 10 |
|
| 11 |
<img src='cheap.png' width='700'>
|
| 12 |
|
|
|
|
|
|
|
|
|
|
| 13 |
[](https://huggingface.co/your-username/your-model-name)
|
| 14 |
[](#license)
|
| 15 |
[](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
|
| 16 |
|
| 17 |
+
**A 4B-parameter Qwen model distilled from Tongyi DeepResearch-30B A3B**, optimized for web-scale “deep research” tasks and inference with **[Alibaba-NLP/DeepResearch](https://github.com/Alibaba-NLP/DeepResearch)**.
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
* **Base**: Qwen 4B (dense)
|
| 20 |
* **Teacher**: Tongyi DeepResearch 30B A3B (MoE)
|
|
|
|
| 34 |
|
| 35 |
* **Primary dataset**: [`cheapresearch/CheapResearch-DS-33k`](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
|
| 36 |
|
|
|
|
| 37 |
---
|
| 38 |
|
| 39 |
## Inference with Alibaba-NLP/DeepResearch (Recommended)
|
|
|
|
| 58 |
MODEL_PATH=cheapresearch/CheapResearch-4B-Thinking
|
| 59 |
```
|
| 60 |
|
|
|
|
|
|
|
| 61 |
### Hardware notes
|
| 62 |
|
| 63 |
+
* **Single 12–16GB GPU** is enough for 4B FP16; FP8/INT4 quantization allows smaller VRAM. If you quantize, the summary model can be local as well.
|
| 64 |
|
| 65 |
---
|
| 66 |
|
| 67 |
## Evaluation
|
| 68 |
|
| 69 |
+
<img src='hle.png' width='700'>
|
| 70 |
+
<img src='simpleqa.png' width='700'>
|
|
|
|
|
|
|
|
|
|
|
|
|
| 71 |
|
| 72 |
## Acknowledgements
|
| 73 |
|