From Context to Skills: Can Language Models Learn from Context Skillfully?
Abstract
A self-evolving framework autonomously discovers and refines context-specific skills for language models through a multi-agent self-play loop with Challenger, Reasoner, and Judge components, improving context learning performance without human supervision.
Many real-world tasks require language models (LMs) to reason over complex contexts that exceed their parametric knowledge. This calls for context learning, where LMs directly learn relevant knowledge from the given context. An intuitive solution is inference-time skill augmentation: extracting the rules and procedures from context into natural-language skills. However, constructing such skills for context learning scenarios faces two challenges: the prohibitive cost of manual skill annotation for long, technically dense contexts, and the lack of external feedback for automated skill construction. In this paper, we propose Ctx2Skill, a self-evolving framework that autonomously discovers, refines, and selects context-specific skills without human supervision or external feedback. At its core, a multi-agent self-play loop has a Challenger that generates probing tasks and rubrics, a Reasoner that attempts to solve them guided by an evolving skill set, and a neutral Judge that provides binary feedback. Crucially, both the Challenger and the Reasoner evolve through accumulated skills: dedicated Proposer and Generator agents analyze failure cases and synthesize them into targeted skill updates for both sides, enabling automated skill discovery and refinement. To prevent adversarial collapse caused by increasingly extreme task generation and over-specialized skill accumulation, we further introduce a Cross-time Replay mechanism that identifies the skill set achieving the best balance across representative cases for the Reasoner side, ensuring robust and generalizable skill evolution. The resulting skills can be plugged into any language model to obtain better context learning capability. Evaluated on four context learning tasks from CL-bench, Ctx2Skill consistently improves solving rates across backbone models.
Community
Ctx2Skill is a self-evolving framework that autonomously discovers, refines, and selects context-specific skills from complex contexts, requiring no human annotation and no external feedback. The resulting natural-language skills can be plugged into any language model at inference time to enhance context learning capability.
ctx2skill is cool in spirit, but i keep coming back to the binary judge as a potential bottleneck that could push the loop into brittle, overfit skills. the cross-time replay trick is clever, yet i worry how representative those historic skills and tasks really are for shifts in context or for out-of-domain data. an ablation where the judge provides richer feedback or where the proposer/generator are regularized to curb task gaming would help, because right now the signal path feels a bit fragile. arxivlens had a solid breakdown that helped me parse the method details; it covers the failure-driven skill edits and cross-time replay nicely, which is worth a look.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Skill Retrieval Augmentation for Agentic AI (2026)
- SkillX: Automatically Constructing Skill Knowledge Bases for Agents (2026)
- SkillLearnBench: Benchmarking Continual Learning Methods for Agent Skill Generation on Real-World Tasks (2026)
- XSkill: Continual Learning from Experience and Skills in Multimodal Agents (2026)
- Training-Free Test-Time Contrastive Learning for Large Language Models (2026)
- OSExpert: Computer-Use Agents Learning Professional Skills via Exploration (2026)
- WebXSkill: Skill Learning for Autonomous Web Agents (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2604.27660 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper