Add library_name, pipeline_tag, and arxiv metadata

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -4,6 +4,10 @@ language:
4
  - en
5
  base_model:
6
  - Qwen/Qwen3-4B
 
 
 
 
7
  metrics:
8
  - accuracy
9
  model-index:
@@ -31,9 +35,9 @@ model-index:
31
 
32
  ![image](fig/scaling.png)
33
 
34
- Recent advancements in Large Language Models (LLMs) have largely focused on depth scaling, where a single agent solves long-horizon problems with multi-turn reasoning and tool use. However, as tasks grow broader, the key bottleneck shifts from individual competence to organizational capability.
35
 
36
- In this work, we explore a complementary dimension of width scaling with multi-agent systems to address broad information seeking. Existing multi-agent systems often rely on hand-crafted workflows and turn-taking interactions that fail to parallelize work effectively. To bridge this gap, we propose WideSeek-R1, a lead-agent-subagent framework trained via multi-agent reinforcement learning (MARL) to synergize scalable orchestration and parallel execution. By utilizing a shared LLM with isolated contexts and specialized tools, WideSeek-R1 jointly optimizes the lead agent and parallel subagents on a curated dataset of 20k broad information-seeking tasks.
37
 
38
  Extensive experiments show that WideSeek-R1-4B achieves an item F1 score of 40.0\% on the WideSearch benchmark, which is comparable to the performance of single-agent DeepSeek-R1-671B. Furthermore, WideSeek-R1-4B exhibits consistent performance gains as the number of parallel subagents increases, highlighting the effectiveness of width scaling.
39
 
@@ -50,4 +54,4 @@ If you use this model in your research, please cite our paper:
50
  journal = {arXiv preprint arXiv:2602.04634},
51
  year = {2026},
52
  }
53
- ```
 
4
  - en
5
  base_model:
6
  - Qwen/Qwen3-4B
7
+ library_name: transformers
8
+ pipeline_tag: text-generation
9
+ tags:
10
+ - arxiv:2602.04634
11
  metrics:
12
  - accuracy
13
  model-index:
 
35
 
36
  ![image](fig/scaling.png)
37
 
38
+ Recent advancements in Large Language Models (LLMs) have largely focused on depth scaling, where a single agent solves long-horizon problems with multi-turn reasoning and tool use. However, as tasks grow broader, the key bottleneck shifts from individual competence to organizational capability.
39
 
40
+ In this work, we explore a complementary dimension of width scaling with multi-agent systems to address broad information seeking. Existing multi-agent systems often rely on hand-crafted workflows and turn-taking interactions that fail to parallelize work effectively. To bridge this gap, we propose WideSeek-R1, a lead-agent-subagent framework trained via multi-agent reinforcement learning (MARL) to synergize scalable orchestration and parallel execution. By utilizing a shared LLM with isolated contexts and specialized tools, WideSeek-R1 jointly optimizes the lead agent and parallel subagents on a curated dataset of 20k broad information-seeking tasks.
41
 
42
  Extensive experiments show that WideSeek-R1-4B achieves an item F1 score of 40.0\% on the WideSearch benchmark, which is comparable to the performance of single-agent DeepSeek-R1-671B. Furthermore, WideSeek-R1-4B exhibits consistent performance gains as the number of parallel subagents increases, highlighting the effectiveness of width scaling.
43
 
 
54
  journal = {arXiv preprint arXiv:2602.04634},
55
  year = {2026},
56
  }
57
+ ```