Safetensors
qwen3
sumo43 commited on
Commit
f9024fe
·
verified ·
1 Parent(s): fea7dfb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -17
README.md CHANGED
@@ -10,17 +10,11 @@ datasets:
10
 
11
  <img src='cheap.png' width='700'>
12
 
13
-
14
- **A 4B-parameter Qwen model distilled from Tongyi DeepResearch-30B A3B**, optimized for web-scale “deep research” tasks and plug-and-play inference with **[Alibaba-NLP/DeepResearch](https://github.com/Alibaba-NLP/DeepResearch)**.
15
-
16
  [![Model](https://img.shields.io/badge/HF-Model-blue)](https://huggingface.co/your-username/your-model-name)
17
  [![License](https://img.shields.io/badge/License-Apache--2.0-green)](#license)
18
  [![Dataset](https://img.shields.io/badge/Dataset-CheapResearch--DS--33k-orange)](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
19
 
20
-
21
- ---
22
-
23
- ## TL;DR
24
 
25
  * **Base**: Qwen 4B (dense)
26
  * **Teacher**: Tongyi DeepResearch 30B A3B (MoE)
@@ -40,7 +34,6 @@ datasets:
40
 
41
  * **Primary dataset**: [`cheapresearch/CheapResearch-DS-33k`](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
42
 
43
-
44
  ---
45
 
46
  ## Inference with Alibaba-NLP/DeepResearch (Recommended)
@@ -65,22 +58,16 @@ Edit the config to add this model
65
  MODEL_PATH=cheapresearch/CheapResearch-4B-Thinking
66
  ```
67
 
68
- > ⚠️ **Note**: Use a **search-enabled** profile in DeepResearch so the model can browse and cite sources. Disable “reasoning suppression” features—this student is trained to produce compact but explicit research traces.
69
-
70
  ### Hardware notes
71
 
72
- * **Single 1624GB GPU** is enough for 4B FP16; FP8/INT4 quantization allows smaller VRAM.
73
 
74
  ---
75
 
76
  ## Evaluation
77
 
78
- | Benchmark | Metric | CheapResearch (4B) | Tongyi DeepResearch (30B A3B) | Notes |
79
- | -------------------- | -------------------: | -----------: | ----------------: | ------------------------------- |
80
- | HLE textonly 200 @1 | Correctness (o4) | — | — | With HLE keyword filtering to prevent cheating |
81
- | SimpleQA @1 | Win-Rate vs Baseline | Correctness (o4) | — | With SimpleQA keyword filtering to prevent cheating |
82
-
83
-
84
 
85
  ## Acknowledgements
86
 
 
10
 
11
  <img src='cheap.png' width='700'>
12
 
 
 
 
13
  [![Model](https://img.shields.io/badge/HF-Model-blue)](https://huggingface.co/your-username/your-model-name)
14
  [![License](https://img.shields.io/badge/License-Apache--2.0-green)](#license)
15
  [![Dataset](https://img.shields.io/badge/Dataset-CheapResearch--DS--33k-orange)](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
16
 
17
+ **A 4B-parameter Qwen model distilled from Tongyi DeepResearch-30B A3B**, optimized for web-scale “deep research” tasks and inference with **[Alibaba-NLP/DeepResearch](https://github.com/Alibaba-NLP/DeepResearch)**.
 
 
 
18
 
19
  * **Base**: Qwen 4B (dense)
20
  * **Teacher**: Tongyi DeepResearch 30B A3B (MoE)
 
34
 
35
  * **Primary dataset**: [`cheapresearch/CheapResearch-DS-33k`](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
36
 
 
37
  ---
38
 
39
  ## Inference with Alibaba-NLP/DeepResearch (Recommended)
 
58
  MODEL_PATH=cheapresearch/CheapResearch-4B-Thinking
59
  ```
60
 
 
 
61
  ### Hardware notes
62
 
63
+ * **Single 1216GB GPU** is enough for 4B FP16; FP8/INT4 quantization allows smaller VRAM. If you quantize, the summary model can be local as well.
64
 
65
  ---
66
 
67
  ## Evaluation
68
 
69
+ <img src='hle.png' width='700'>
70
+ <img src='simpleqa.png' width='700'>
 
 
 
 
71
 
72
  ## Acknowledgements
73