Cursor trained Composer to self-summarize through RL

What's new? Composer is a new AI model by Cursor for agentic coding using self-summarization for long horizon tasks; it compacts code context with internal tests;

· 1 min read
Cursor

Composer, developed by Cursor, is a new AI model tailored for agentic coding, specifically designed to address the challenges of long-horizon tasks. This model incorporates a reinforcement learning training process called self-summarization, which enables it to retain and process information across much longer sequences than traditional context window limits. The release targets developers and researchers who require AI to assist with complex coding tasks that demand sustained reasoning and context retention over hundreds of steps.

Currently, Composer is evaluated internally using CursorBench and Terminal-Bench 2.0, with performance metrics indicating higher accuracy and efficiency in token usage compared to established prompt-based compaction methods. The model achieves compaction with summaries averaging 1,000 tokens, significantly reducing context loss and computational overhead.

Cursor

Cursor, the company behind Composer, is focused on advancing agentic AI capabilities in software engineering scenarios. The company's approach integrates compaction directly into the training loop, allowing Composer to learn which contextual information is most critical to retain. By leveraging reinforcement learning rewards for effective self-summaries, Cursor positions Composer as a robust alternative to current compaction strategies, which often rely on lengthy, manually crafted prompts and are susceptible to information loss.

Early testing shows Composer solving problems that have stumped other large models, illustrating its potential for tackling more complex, multi-step programming challenges in future releases.

Source