Forget Fine-Tuning — ACE Lets LLMs Self-Improve by Evolving Context Playbooks

Overview

Agentic Context Engineering (ACE) reframes adaptation as a context-first process: instead of updating model weights, ACE incrementally edits and expands a persistent input ‘playbook’ that captures domain tactics and lessons. The framework was evaluated by researchers from Stanford, SambaNova Systems and UC Berkeley and shows notable gains on agent benchmarks and finance reasoning while dramatically reducing adaptation latency and token costs.

How ACE works

ACE treats context as a living artifact maintained by three roles that use the same base LLM to isolate context effects:

Design choices such as small incremental delta updates and a grow-and-refine workflow preserve useful history and avoid ‘context collapse’ that can happen with monolithic rewrites. The team fixed a single base model (non-thinking DeepSeek-V3.1) across roles to measure pure context effects.

Benchmarks and results

AppWorld agents

Built on the ReAct baseline, ReAct+ACE outperformed strong context-adaptation baselines including ICL, GEPA and Dynamic Cheatsheet. Reported improvements include +10.6% average over selected baselines and ~+7.6% over Dynamic Cheatsheet in online adaptation. On a Sept 20, 2025 snapshot of the AppWorld leaderboard, ReAct+ACE reached 59.4% versus IBM CUGA at 60.3% (GPT-4.1), while using a smaller open-source base model; ACE even surpassed CUGA on the more difficult test-challenge split.

Finance reasoning

On FiNER token tagging and XBRL Formula numerical reasoning, ACE achieved an average +8.6% over baselines when offline ground-truth labels were available. The method also works with execution-only feedback, though the quality of feedback signals affects final performance.

Cost, latency and efficiency

ACE minimizes adaptation overhead by using non-LLM merges and localized updates. Measured reductions include offline AppWorld latency drops around −82.3% and −75.1% fewer rollouts versus GEPA. In online FiNER scenarios, ACE reported −91.5% latency and −83.6% token cost compared to Dynamic Cheatsheet. These savings stem from deterministic, small delta merges and targeted playbook growth instead of repeated heavy re-generation.

Key implications

ACE positions ‘context engineering’ as a practical alternative to weight updates for many agentic tasks. By maintaining a curated, incrementally growing playbook of tactics, LLM-based agents can self-tune using the same model weights while preserving useful histories and avoiding destructive rewrites. The approach delivers measurable accuracy gains on agent and finance benchmarks and cuts adaptation latency and token costs substantially.

Limitations and considerations

ACE’s benefits depend on the quality of feedback signals and the complexity of tasks. Deterministic merges and pruning help, but outcomes still track the fidelity of Reflector summaries and the strength of signal provided to Curator. In production, teams will need to monitor playbook growth and feedback pipelines to avoid encoding noisy or biased lessons.

Where to find the paper

Full details, experiments and analyses are available in the ACE paper: https://arxiv.org/pdf/2510.04618