code-pruner / README.md
ayanami-kitasan's picture
Update README.md
863cee3 verified
---
license: mit
library_name: transformers
base_model: Qwen/Qwen3-Reranker-0.6B
pipeline_tag: token-classification
tags:
- code
- context-pruning
- agent
datasets:
- nick007x/github-code-2025
metrics:
- f1
- mse
---
# SWE-Pruner: Self-Adaptive Context Pruning for Coding Agents
SWE-Pruner is a self-adaptive context pruning framework specifically designed for coding agents. It addresses the challenges of long interaction contexts, such as high API costs and latency, by performing task-aware adaptive pruning.
- **Paper:** [SWE-Pruner: Self-Adaptive Context Pruning for Coding Agents](https://huggingface.co/papers/2601.16746)
- **Repository:** [https://github.com/Ayanami1314/swe-pruner](https://github.com/Ayanami1314/swe-pruner)
## Description
Inspired by how human programmers selectively skim code, SWE-Pruner enables agents to formulate explicit goals (e.g., "focus on error handling") which guide a lightweight neural skimmer (0.6B parameters). This skimmer dynamically selects relevant lines from the surrounding context, preserving critical implementation details while significantly reducing token usage.
Evaluations across benchmarks show that SWE-Pruner achieves 23-54% token reduction on agent tasks like SWE-Bench Verified and up to 14.84x compression on single-turn tasks like LongCodeQA with minimal performance impact.
## Model Usage
Given that we have made significant modifications to the model, its dual-head architecture and the complex compression head service code will be rather complex.
Therefore, we recommend that you use the version we have released on [GitHub](https://github.com/Ayanami1314/swe-pruner) instead of attempting to use the original model on your own.
## Citation
If you find SWE-Pruner useful in your research, please cite:
```bibtex
@misc{wang2026sweprunerselfadaptivecontextpruning,
title={SWE-Pruner: Self-Adaptive Context Pruning for Coding Agents},
author={Yuhang Wang and Yuling Shi and Mo Yang and Rongrui Zhang and Shilin He and Heng Lian and Yuting Chen and Siyu Ye and Kai Cai and Xiaodong Gu},
year={2026},
eprint={2601.16746},
archivePrefix={arXiv},
primaryClass={cs.SE},
url={https://arxiv.org/abs/2601.16746},
}
```