Safetensors
qwen3
sumo43 commited on
Commit
f483428
·
verified ·
1 Parent(s): 09f3130

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +152 -0
README.md ADDED
@@ -0,0 +1,152 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - cheapresearch/CheapResearch-DS-33k
5
+ ---
6
+
7
+
8
+
9
+ # CheapResearch-4B-Thinking
10
+
11
+ > **A 4B-parameter Qwen model distilled from Tongyi DeepResearch-30B A3B**, optimized for web-scale “deep research” tasks and plug-and-play inference with **[Alibaba-NLP/DeepResearch](https://github.com/Alibaba-NLP/DeepResearch)**.
12
+
13
+ [![Model](https://img.shields.io/badge/HF-Model-blue)](https://huggingface.co/your-username/your-model-name)
14
+ [![License](https://img.shields.io/badge/License-Apache--2.0-green)](#license)
15
+ [![Dataset](https://img.shields.io/badge/Dataset-CheapResearch--DS--33k-orange)](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
16
+
17
+ ---
18
+
19
+ ## TL;DR
20
+
21
+ * **Base**: Qwen 4B (dense)
22
+ * **Teacher**: Tongyi DeepResearch 30B A3B (MoE)
23
+ * **Method**: SFT distillation on **33k** curated deep-research examples
24
+ * **Dataset**: [`cheapresearch/CheapResearch-DS-33k`](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
25
+ * **Primary Use**: Fast, low-cost **DeepResearch** agent runs (browsing, multi-step reasoning, source-grounded answers)
26
+
27
+ ---
28
+
29
+ ### Intended Use
30
+
31
+ * Browser-based local research assistant (via **Alibaba-NLP/DeepResearch**)
32
+ * Low-latency DR on modest GPUs/CPUs
33
+
34
+
35
+ ## Training Data
36
+
37
+ * **Primary dataset**: [`cheapresearch/CheapResearch-DS-33k`](https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k)
38
+
39
+
40
+ ---
41
+
42
+ ## Inference with Alibaba-NLP/DeepResearch (Recommended)
43
+
44
+ This model is intended to be used **directly** with the DeepResearch repo.
45
+
46
+ ### 1) Install & set up
47
+
48
+ ```bash
49
+ git clone https://github.com/Alibaba-NLP/DeepResearch
50
+ cd DeepResearch
51
+ # Create env (example)
52
+ python -m venv .venv && source .venv/bin/activate
53
+ pip install -e . # or pip install -r requirements.txt if provided
54
+ ```
55
+
56
+ ### 2) Point DeepResearch to this model
57
+
58
+ Edit the config to add this model
59
+
60
+ ```bash
61
+ MODEL_PATH=cheapresearch/CheapResearch-4B-Thinking
62
+ ```
63
+
64
+ > ⚠️ **Note**: Use a **search-enabled** profile in DeepResearch so the model can browse and cite sources. Disable “reasoning suppression” features—this student is trained to produce compact but explicit research traces.
65
+
66
+ ### Hardware notes
67
+
68
+ * **Single 16–24GB GPU** is enough for 4B FP16; FP8/INT4 quantization allows smaller VRAM.
69
+
70
+ ---
71
+
72
+ ## Evaluation
73
+
74
+ | Benchmark | Metric | CheapResearch (4B) | Tongyi DeepResearch (30B A3B) | Notes |
75
+ | -------------------- | -------------------: | -----------: | ----------------: | ------------------------------- |
76
+ | HLE textonly 200 @1 | Correctness (o4) | — | — | With HLE keyword filtering to prevent cheating |
77
+ | SimpleQA @1 | Win-Rate vs Baseline | Correctness (o4) | — | With SimpleQA keyword filtering to prevent cheating |
78
+
79
+
80
+
81
+ ## Acknowledgements
82
+
83
+ * Qwen team for the base 4B architecture
84
+ * Alibaba-NLP for **DeepResearch**
85
+ * CheapResearch contributors for the 33k dataset
86
+
87
+ ---
88
+
89
+ ## Citation
90
+
91
+ If you use this model, please cite:
92
+
93
+ ```bibtex
94
+ @software{qwen4b_deepresearch_distill_2025,
95
+ title = {Qwen-4B DeepResearch-Distill},
96
+ author = {Your Name},
97
+ year = {2025},
98
+ url = {https://huggingface.co/your-username/your-model-name}
99
+ }
100
+ ```
101
+
102
+ And the dataset:
103
+
104
+ ```bibtex
105
+ @dataset{cheapresearch_ds_33k,
106
+ title = {CheapResearch-DS-33k},
107
+ author = {CheapResearch Contributors},
108
+ year = {2025},
109
+ url = {https://huggingface.co/datasets/cheapresearch/CheapResearch-DS-33k}
110
+ }
111
+ ```
112
+
113
+ ---
114
+
115
+ ## Changelog
116
+
117
+ * **v1.0.0 (2025-10-03)** — First public release (33k distillation, DeepResearch-ready)
118
+
119
+ ---
120
+
121
+ ### Model Card Metadata (Hugging Face)
122
+
123
+ ```yaml
124
+ ---
125
+ language:
126
+ - en
127
+ license: apache-2.0
128
+ library_name: transformers
129
+ pipeline_tag: text-generation
130
+ tags:
131
+ - qwen
132
+ - deep-research
133
+ - browsing
134
+ - citation
135
+ - reasoning
136
+ - distillation
137
+ - agent
138
+ - vllm
139
+ - cheapresearch
140
+ datasets:
141
+ - cheapresearch/CheapResearch-DS-33k
142
+ base_model:
143
+ - Qwen/Qwen2.5-4B
144
+ model-index:
145
+ - name: Qwen-4B DeepResearch-Distill
146
+ results: []
147
+ ---
148
+ ```
149
+
150
+ ---
151
+
152
+ If you share your **actual model ID** and any concrete eval numbers or training hyperparams, I’ll slot them in and tighten the “Training Procedure” and “Evaluation” sections for you.