ai · 5 min read · May 1, 2026

Self-Evolving Skills Let Language Models Learn From Long Context

Ctx2Skill uses multi-agent loops to automatically extract and refine skills from dense context without human annotation or external feedback.

Source: arxiv/cs.AI · Shuzheng Si, Haozhe Zhao, Yu Lei, Qingyi Wang, Dingwei Chen, Zhitong Wang, Zhenhailong Wang, Kangyang Luo, Zheng Wang, Gang Chen, Fanchao Qi, Minjia Zhang, Maosong Sun · open original ↗

A framework autonomously discovers and refines natural-language skills from complex context to improve language model reasoning without manual supervision.

  • Language models struggle with reasoning over long, dense contexts beyond their training knowledge.
  • Manual skill annotation is expensive; automated skill construction lacks feedback signals.
  • Ctx2Skill uses three agents: Challenger generates probing tasks, Reasoner solves them, Judge provides binary feedback.
  • Proposer and Generator agents analyze failures and synthesize skill updates for both Challenger and Reasoner.
  • Cross-time Replay mechanism prevents adversarial collapse by selecting balanced skill sets.
  • Extracted skills plug into any language model to enhance context learning performance.
  • Tested on CL-bench tasks; shows consistent improvement across different backbone models.

Astrobobo tool mapping

  • Knowledge Capture Record procedural rules and decision rubrics from your domain as structured skill definitions. Use these as prompts to guide language model reasoning on context-heavy tasks.
  • Focus Brief Before running a context-learning task, summarize the key skills and constraints the model should apply. This mirrors the Reasoner's guided reasoning in Ctx2Skill.
  • Reading Queue Queue domain-specific documentation (technical specs, policies, case studies) and extract reusable skills from failure cases as you process them.

Frequently asked

  • Context learning means a language model must reason over information provided in the input (e.g., a long document or technical specification) that is not in its training data. Models struggle because they lack explicit procedures to extract and apply rules from dense, unfamiliar contexts. Ctx2Skill addresses this by automatically discovering skills—natural-language procedures—that guide the model to reason over new context correctly.
Share X LinkedIn
cite
APA
Shuzheng Si, Haozhe Zhao, Yu Lei, Qingyi Wang, Dingwei Chen, Zhitong Wang, Zhenhailong Wang, Kangyang Luo, Zheng Wang, Gang Chen, Fanchao Qi, Minjia Zhang, Maosong Sun. (2026, May 1). Self-Evolving Skills Let Language Models Learn From Long Context. Astrobobo Content Engine (rewrite of arxiv/cs.AI). https://astrobobo-content-engine.vercel.app/article/self-evolving-skills-let-language-models-learn-from-long-context-4b88f5
MLA
Shuzheng Si, Haozhe Zhao, Yu Lei, Qingyi Wang, Dingwei Chen, Zhitong Wang, Zhenhailong Wang, Kangyang Luo, Zheng Wang, Gang Chen, Fanchao Qi, Minjia Zhang, Maosong Sun. "Self-Evolving Skills Let Language Models Learn From Long Context." Astrobobo Content Engine, 1 May 2026, https://astrobobo-content-engine.vercel.app/article/self-evolving-skills-let-language-models-learn-from-long-context-4b88f5. Based on "arxiv/cs.AI", https://arxiv.org/abs/2604.27660.
BibTeX
@misc{astrobobo_self-evolving-skills-let-language-models-learn-from-long-context-4b88f5_2026,
  author       = {Shuzheng Si, Haozhe Zhao, Yu Lei, Qingyi Wang, Dingwei Chen, Zhitong Wang, Zhenhailong Wang, Kangyang Luo, Zheng Wang, Gang Chen, Fanchao Qi, Minjia Zhang, Maosong Sun},
  title        = {Self-Evolving Skills Let Language Models Learn From Long Context},
  year         = {2026},
  url          = {https://astrobobo-content-engine.vercel.app/article/self-evolving-skills-let-language-models-learn-from-long-context-4b88f5},
  note         = {Astrobobo rewrite of arxiv/cs.AI, https://arxiv.org/abs/2604.27660},
}

Related insights