您正在查看静态缓存页面 · 查看完整动态版本 · 登录 参与讨论
递归语言模型的无限回响:当AI学会“翻书”而非“死记硬背”
✨步子哥 (steper) 话题创建于 2026-01-07 15:15:57
回复 #1
✨步子哥 (steper)
2026年01月08日 00:02
Recursive Language Models Poster

Recursive Language Models

Scaling AI Beyond Context Windows

Alex L. Zhang, Tim Kraska, Omar Khattab (MIT CSAIL) • 2025
The Paradigm Shift

Traditional LLMs struggle with Context Rot — performance degradation when inputs exceed standard context windows (e.g., GPT-5's 262K tokens).

Recursive Language Models (RLMs) propose a new inference paradigm. Instead of cramming prompts into a context window, RLMs treat the input as a programmable environment (Python REPL). This enables LLMs to interact with massive inputs symbolically via code, handling up to 10 Million tokens while improving accuracy and reducing costs.

How RLMs Work

The REPL-based recursive workflow:

1
Store Context: Full input loaded as a variable (e.g., context) in Python REPL.
2
Generate Code: Root LLM writes code to slice, filter, or search the context.
3
Recursive Calls: Sub-LLMs invoked on relevant subsets (llm_query).
4
Aggregate: Results collected and final answer output via FINAL().
Key Benefits
📏
Scalability
Handles inputs 100x larger than standard limits (10M+ tokens).
🎯
Accuracy
Double-digit improvements on long-context benchmarks.
💰
Cost Efficiency
Median costs equal to or lower than base models.
Reduced Hallucination
Self-verification via code execution and iterative refinement.
Performance Benchmarks

Comparison on GPT-5 across long-context tasks. RLMs significantly outperform baselines.

Method / Model S-NIAH (%) BrowseComp+ (%) OOLONG (%) Avg. Cost ($)
Base GPT-5 Fails (>262K) 0.00 12.50 N/A
Summary Agent 85.00 45.67 34.00 8.98
CodeAct + BM25 78.00 52.33 41.00 5.12
RLM (GPT-5) 92.00 91.33 56.50 0.99
Applications & Future Outlook
🚀 Long-Horizon Agents
Enables autonomous agents to operate over massive document sets and codebases without memory loss.
📚 Document Analysis
Perfect for semantic aggregation, multi-hop QA, and deep search in legal, medical, and financial texts.
🔓 Open Source Ecosystem
Available on GitHub (alexzhang13/rlm). Integrated with Prime Intellect for parallelization and RL training.
arXiv:2512.24601 • MIT CSAIL • github.com/alexzhang13/rlm
Designed for AI Researchers & Engineers