Add pipeline tag, library name and sample usage
#1
by nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,10 +1,13 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
-
datasets:
|
| 4 |
-
- OpenResearcher/OpenResearcher-Dataset
|
| 5 |
base_model:
|
| 6 |
- nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-Base-BF16
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
---
|
|
|
|
| 8 |
<div style="display: flex; align-items: center; justify-content: center; gap: 8px;">
|
| 9 |
<img src="https://raw.githubusercontent.com/TIGER-AI-Lab/OpenResearcher/main/assets/imgs/or-logo1.png" style="height: 84px; width: auto;">
|
| 10 |
<img src="https://raw.githubusercontent.com/TIGER-AI-Lab/OpenResearcher/refs/heads/main/assets/imgs/openresearcher-title.svg" style="height: 84px; width: auto;">
|
|
@@ -13,17 +16,12 @@ base_model:
|
|
| 13 |
<div align="center">
|
| 14 |
<a href="https://arxiv.org/abs/2603.20278"><img src="https://img.shields.io/badge/arXiv-B31B1B?style=for-the-badge&logo=arXiv&logoColor=white" alt="Blog"></a>
|
| 15 |
<a href="https://huggingface.co/papers/2603.20278"><img src="https://img.shields.io/badge/Paper-FFD966?style=for-the-badge&logo=huggingface&logoColor=ffffff" alt="Model"></a>
|
| 16 |
-
<
|
| 17 |
-
<!-- <a href="https://x.com/DongfuJiang/status/2020946549422031040"><img src="https://img.shields.io/badge/Twitter-000000?style=for-the-badge&logo=X&logoColor=white" alt="Blog"></a> -->
|
| 18 |
-
<!-- <a href="https://boiled-honeycup-4c7.notion.site/OpenResearcher-A-Fully-Open-Pipeline-for-Long-Horizon-Deep-Research-Trajectory-Synthesis-2f7e290627b5800cb3a0cd7e8d6ec0ea?source=copy_link"><img src="https://img.shields.io/badge/Blog-4285F4?style=for-the-badge&logo=google-chrome&logoColor=white" alt="Blog"></a> -->
|
| 19 |
-
<a href="https://github.com/TIGER-AI-Lab/OpenResearcher"><img src="https://img.shields.io/badge/Github-181717?style=for-the-badge&logo=github&logoColor=white" alt="Blog"></a>
|
| 20 |
<a href="https://huggingface.co/datasets/OpenResearcher/OpenResearcher-Dataset"><img src="https://img.shields.io/badge/Dataset-FFB7B2?style=for-the-badge&logo=huggingface&logoColor=ffffff" alt="Dataset"></a>
|
| 21 |
<a href="https://huggingface.co/OpenResearcher/Nemotron-3-Nano-30B-A3B"><img src="https://img.shields.io/badge/Model-FFD966?style=for-the-badge&logo=huggingface&logoColor=ffffff" alt="Model"></a>
|
| 22 |
<a href="https://huggingface.co/spaces/OpenResearcher/OpenResearcher"><img src="https://img.shields.io/badge/Demo-F97316.svg?style=for-the-badge&logo=gradio&logoColor=white" alt="Demo"></a>
|
| 23 |
-
<!-- <a href="https://wandb.ai/dongfu/nano-v3-sft-search"><img src="https://img.shields.io/badge/WandB%20Logs-48B5A3?style=for-the-badge&logo=weightsandbiases&logoColor=white" alt="WandB Logs"></a> -->
|
| 24 |
<a href="https://huggingface.co/datasets/OpenResearcher/OpenResearcher-Eval-Logs/tree/main"><img src="https://img.shields.io/badge/Eval%20Logs-755BB4?style=for-the-badge&logo=google-sheets&logoColor=white" alt="Eval Logs"></a>
|
| 25 |
</div>
|
| 26 |
-
</div>
|
| 27 |
|
| 28 |
<div align="center" style="padding: 10px 0 -4px; display: flex; align-items: center; justify-content: center; gap: 16px;">
|
| 29 |
<div style="width: 60px; height: 2px; background: linear-gradient(90deg, transparent, #E24B4A);"></div>
|
|
@@ -46,7 +44,39 @@ base_model:
|
|
| 46 |
</p>
|
| 47 |
|
| 48 |
## Overview
|
| 49 |
-
**OpenResearcher** is a fully open agentic large language model (30B-A3B) designed for **long-horizon deep research** scenarios. It achieves an impressive **54.8%** accuracy on [BrowseComp-Plus](https://huggingface.co/spaces/Tevatron/BrowseComp-Plus), surpassing performance of `GPT-4.1`, `Claude-Opus-4`, `Gemini-2.5-Pro`, `DeepSeek-R1` and `Tongyi-DeepResearch`.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
|
| 51 |
## OpenResearcher-30B-A3B-GGUF
|
| 52 |
**Note: For the best performance, we recommend using [OpenResearcher-30B-A3B](https://huggingface.co/OpenResearcher/OpenResearcher-30B-A3B).**
|
|
@@ -70,5 +100,4 @@ To support efficient deployment, we release several quantized versions of [OpenR
|
|
| 70 |
journal={arXiv preprint arXiv:2603.20278},
|
| 71 |
year={2026}
|
| 72 |
}
|
| 73 |
-
```
|
| 74 |
-
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-Base-BF16
|
| 4 |
+
datasets:
|
| 5 |
+
- OpenResearcher/OpenResearcher-Dataset
|
| 6 |
+
license: mit
|
| 7 |
+
pipeline_tag: text-generation
|
| 8 |
+
library_name: llama.cpp
|
| 9 |
---
|
| 10 |
+
|
| 11 |
<div style="display: flex; align-items: center; justify-content: center; gap: 8px;">
|
| 12 |
<img src="https://raw.githubusercontent.com/TIGER-AI-Lab/OpenResearcher/main/assets/imgs/or-logo1.png" style="height: 84px; width: auto;">
|
| 13 |
<img src="https://raw.githubusercontent.com/TIGER-AI-Lab/OpenResearcher/refs/heads/main/assets/imgs/openresearcher-title.svg" style="height: 84px; width: auto;">
|
|
|
|
| 16 |
<div align="center">
|
| 17 |
<a href="https://arxiv.org/abs/2603.20278"><img src="https://img.shields.io/badge/arXiv-B31B1B?style=for-the-badge&logo=arXiv&logoColor=white" alt="Blog"></a>
|
| 18 |
<a href="https://huggingface.co/papers/2603.20278"><img src="https://img.shields.io/badge/Paper-FFD966?style=for-the-badge&logo=huggingface&logoColor=ffffff" alt="Model"></a>
|
| 19 |
+
<a href="https://github.com/TIGER-AI-Lab/OpenResearcher"><img src="https://img.shields.io/badge/Github-181717?style=for-the-badge&logo=github&logoColor=white" alt="Blog"></a>
|
|
|
|
|
|
|
|
|
|
| 20 |
<a href="https://huggingface.co/datasets/OpenResearcher/OpenResearcher-Dataset"><img src="https://img.shields.io/badge/Dataset-FFB7B2?style=for-the-badge&logo=huggingface&logoColor=ffffff" alt="Dataset"></a>
|
| 21 |
<a href="https://huggingface.co/OpenResearcher/Nemotron-3-Nano-30B-A3B"><img src="https://img.shields.io/badge/Model-FFD966?style=for-the-badge&logo=huggingface&logoColor=ffffff" alt="Model"></a>
|
| 22 |
<a href="https://huggingface.co/spaces/OpenResearcher/OpenResearcher"><img src="https://img.shields.io/badge/Demo-F97316.svg?style=for-the-badge&logo=gradio&logoColor=white" alt="Demo"></a>
|
|
|
|
| 23 |
<a href="https://huggingface.co/datasets/OpenResearcher/OpenResearcher-Eval-Logs/tree/main"><img src="https://img.shields.io/badge/Eval%20Logs-755BB4?style=for-the-badge&logo=google-sheets&logoColor=white" alt="Eval Logs"></a>
|
| 24 |
</div>
|
|
|
|
| 25 |
|
| 26 |
<div align="center" style="padding: 10px 0 -4px; display: flex; align-items: center; justify-content: center; gap: 16px;">
|
| 27 |
<div style="width: 60px; height: 2px; background: linear-gradient(90deg, transparent, #E24B4A);"></div>
|
|
|
|
| 44 |
</p>
|
| 45 |
|
| 46 |
## Overview
|
| 47 |
+
**OpenResearcher** is a fully open agentic large language model (30B-A3B) designed for **long-horizon deep research** scenarios, presented in [OpenResearcher: A Fully Open Pipeline for Long-Horizon Deep Research Trajectory Synthesis](https://huggingface.co/papers/2603.20278). It achieves an impressive **54.8%** accuracy on [BrowseComp-Plus](https://huggingface.co/spaces/Tevatron/BrowseComp-Plus), surpassing performance of `GPT-4.1`, `Claude-Opus-4`, `Gemini-2.5-Pro`, `DeepSeek-R1` and `Tongyi-DeepResearch`.
|
| 48 |
+
|
| 49 |
+
## Sample Usage
|
| 50 |
+
|
| 51 |
+
The following example demonstrates how to run a research task using the OpenResearcher pipeline. Note that this requires deploying the model (e.g., via vLLM) and configuring the environment as described in the [official repository](https://github.com/TIGER-AI-Lab/OpenResearcher).
|
| 52 |
+
|
| 53 |
+
```python
|
| 54 |
+
import asyncio
|
| 55 |
+
from deploy_agent import run_one, BrowserPool
|
| 56 |
+
from utils.openai_generator import OpenAIAsyncGenerator
|
| 57 |
+
|
| 58 |
+
async def main():
|
| 59 |
+
# Initialize generator and browser
|
| 60 |
+
generator = OpenAIAsyncGenerator(
|
| 61 |
+
base_url="http://localhost:8001/v1",
|
| 62 |
+
model_name="OpenResearcher/OpenResearcher-30B-A3B",
|
| 63 |
+
use_native_tools=True
|
| 64 |
+
)
|
| 65 |
+
browser_pool = BrowserPool(search_url=None, browser_backend="serper")
|
| 66 |
+
|
| 67 |
+
# Run deep research
|
| 68 |
+
await run_one(
|
| 69 |
+
question="What is the latest news about OpenAI?",
|
| 70 |
+
qid="quick_start",
|
| 71 |
+
generator=generator,
|
| 72 |
+
browser_pool=browser_pool,
|
| 73 |
+
)
|
| 74 |
+
|
| 75 |
+
browser_pool.cleanup("quick_start")
|
| 76 |
+
|
| 77 |
+
if __name__ == "__main__":
|
| 78 |
+
asyncio.run(main())
|
| 79 |
+
```
|
| 80 |
|
| 81 |
## OpenResearcher-30B-A3B-GGUF
|
| 82 |
**Note: For the best performance, we recommend using [OpenResearcher-30B-A3B](https://huggingface.co/OpenResearcher/OpenResearcher-30B-A3B).**
|
|
|
|
| 100 |
journal={arXiv preprint arXiv:2603.20278},
|
| 101 |
year={2026}
|
| 102 |
}
|
| 103 |
+
```
|
|
|