ZhuofengLi commited on
Commit
4ce1b07
·
verified ·
1 Parent(s): dc8f587

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -25,7 +25,7 @@ base_model:
25
  </p>
26
 
27
  ## OpenResearcher-30B-A3B Overview
28
- OpenResearcher-30B-A3B is an agentic large language model designed for long-horizon deep research fine-tuned from [NVIDIA-Nemotron-3-Nano-30B-A3B-Base-BF16](https://huggingface.co/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-Base-BF16) on 96K [OpenResearcher dataset](https://huggingface.co/datasets/OpenResearcher/OpenResearcher-Dataset). The dataset is derived by distilling GPT-OSS-120B with [native browser tools](https://docs.vllm.ai/projects/recipes/en/latest/OpenAI/GPT-OSS.html#usage:~:text=Limitation%20section%20below.-,Tool%20Use,-%C2%B6). More info about the dataset can be found on the dataset card at [OpenResearcher dataset](https://huggingface.co/datasets/OpenResearcher/OpenResearcher-Dataset).
29
 
30
  The model achieves an impressive **54.8%** accuracy on [BrowseComp-Plus](https://huggingface.co/spaces/Tevatron/BrowseComp-Plus), surpassing performance of `GPT-4.1`, `Claude-Opus-4`, `Gemini-2.5-Pro`, `DeepSeek-R1` and `Tongyi-DeepResearch`.
31
  <div align="center">
 
25
  </p>
26
 
27
  ## OpenResearcher-30B-A3B Overview
28
+ OpenResearcher-30B-A3B is an agentic large language model designed for long-horizon deep research fine-tuned from [NVIDIA-Nemotron-3-Nano-30B-A3B-Base-BF16](https://huggingface.co/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-Base-BF16) on 96K [OpenResearcher dataset](https://huggingface.co/datasets/OpenResearcher/OpenResearcher-Dataset) with **100+** turns. The dataset is derived by distilling GPT-OSS-120B with [native browser tools](https://docs.vllm.ai/projects/recipes/en/latest/OpenAI/GPT-OSS.html#usage:~:text=Limitation%20section%20below.-,Tool%20Use,-%C2%B6). More info about the dataset can be found on the dataset card at [OpenResearcher dataset](https://huggingface.co/datasets/OpenResearcher/OpenResearcher-Dataset).
29
 
30
  The model achieves an impressive **54.8%** accuracy on [BrowseComp-Plus](https://huggingface.co/spaces/Tevatron/BrowseComp-Plus), surpassing performance of `GPT-4.1`, `Claude-Opus-4`, `Gemini-2.5-Pro`, `DeepSeek-R1` and `Tongyi-DeepResearch`.
31
  <div align="center">