Home / Technology / MIT's RLMs Unlock Millions of Tokens
MIT's RLMs Unlock Millions of Tokens
21 Jan
Summary
- RLMs treat prompts as external environments for LLMs to code against.
- The technique processes millions of tokens without retraining models.
- RLMs show significant performance gains on large-scale benchmarks.

Researchers at MIT CSAIL have introduced Recursive Language Models (RLMs), a novel inference technique that redefines how large language models (LLMs) handle extensive prompts. Instead of fitting entire texts into a model's context window, RLMs enable LLMs to programmatically interact with prompts as external environments. This approach allows models to decompose and recursively process text snippets, effectively reasoning over millions of tokens without the need for retraining.
This framework reframes long-context reasoning as a systems problem, offering enterprises a viable solution for complex tasks such as codebase analysis and legal review. By acting as a wrapper around existing models, RLMs can be seamlessly integrated into current applications. The method draws inspiration from classical computing's 'out-of-core' algorithms, loading text as a variable that the LLM then manipulates using code to extract and analyze relevant chunks.




