nielsr HF Staff commited on
Commit
41cfcf3
·
verified ·
1 Parent(s): c5a550f

Add model card and metadata

Browse files

Hi! I'm Niels from the Hugging Face community team. I noticed that this repository didn't have a model card, so I've opened this PR to add one. It includes relevant metadata like the pipeline tag, library name, and base model, as well as links to the paper and GitHub repository to make the model more discoverable.

Files changed (1) hide show
  1. README.md +35 -3
README.md CHANGED
@@ -1,3 +1,35 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ library_name: transformers
4
+ base_model: Qwen/Qwen3-Reranker-0.6B
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - code
8
+ - context-pruning
9
+ ---
10
+
11
+ # SWE-Pruner: Self-Adaptive Context Pruning for Coding Agents
12
+
13
+ SWE-Pruner is a self-adaptive context pruning framework specifically designed for coding agents. It addresses the challenges of long interaction contexts, such as high API costs and latency, by performing task-aware adaptive pruning.
14
+
15
+ - **Paper:** [SWE-Pruner: Self-Adaptive Context Pruning for Coding Agents](https://huggingface.co/papers/2601.16746)
16
+ - **Repository:** [https://github.com/Ayanami1314/swe-pruner](https://github.com/Ayanami1314/swe-pruner)
17
+
18
+ ## Description
19
+ Inspired by how human programmers selectively skim code, SWE-Pruner enables agents to formulate explicit goals (e.g., "focus on error handling") which guide a lightweight neural skimmer (0.6B parameters). This skimmer dynamically selects relevant lines from the surrounding context, preserving critical implementation details while significantly reducing token usage.
20
+
21
+ Evaluations across benchmarks show that SWE-Pruner achieves 23-54% token reduction on agent tasks like SWE-Bench Verified and up to 14.84x compression on single-turn tasks like LongCodeQA with minimal performance impact.
22
+
23
+ ## Citation
24
+ If you find SWE-Pruner useful in your research, please cite:
25
+ ```bibtex
26
+ @misc{wang2026sweprunerselfadaptivecontextpruning,
27
+ title={SWE-Pruner: Self-Adaptive Context Pruning for Coding Agents},
28
+ author={Yuhang Wang and Yuling Shi and Mo Yang and Rongrui Zhang and Shilin He and Heng Lian and Yuting Chen and Siyu Ye and Kai Cai and Xiaodong Gu},
29
+ year={2026},
30
+ eprint={2601.16746},
31
+ archivePrefix={arXiv},
32
+ primaryClass={cs.SE},
33
+ url={https://arxiv.org/abs/2601.16746},
34
+ }
35
+ ```