ShinkaEvolve: LLM-Driven Program Evolution Reaches SOTA Circle Packing in ~150 Evaluations

What ShinkaEvolve aims to do

ShinkaEvolve is an open-source framework from Sakana AI that couples large language models (LLMs) with evolutionary search to evolve programs for scientific and engineering problems. The core claim is dramatic sample efficiency: instead of thousands of program evaluations, ShinkaEvolve can find strong solutions with only hundreds of evaluations in several benchmark domains.

Key techniques that cut evaluation costs

The framework reduces wasted evaluations with three interacting mechanisms:

Benchmarks and empirical results

ShinkaEvolve was evaluated across four domains and showed consistent gains under tight evaluation budgets:

How the evolutionary loop operates in practice

ShinkaEvolve maintains an archive of evaluated programs with fitness, public metrics, and textual feedback. Each generation follows these steps:

  1. Sample an island and select parent(s) according to adaptive policies.
  2. Build a mutation context combining top-K candidates and random “inspiration” programs.
  3. Propose edits via three operators: diff edits, full rewrites, and LLM-guided crossovers, while protecting immutable code regions with explicit markers.
  4. Apply novelty filtering on proposed candidates and run only those passing the judge.
  5. Execute evaluated candidates, update the archive, and update bandit statistics that steer future LLM selection.

The system also periodically generates a meta-scratchpad summarizing recently successful strategies; those summaries are fed back into prompts to accelerate later generations.

Concrete engineering discoveries

ShinkaEvolve did not simply reapply hand-coded strategies. Examples of discovered techniques include:

Comparison with prior systems

Closed-source systems like AlphaEvolve reported strong results but relied on far larger evaluation budgets. ShinkaEvolve reproduces and surpasses the circle-packing result using orders-of-magnitude fewer samples, and it releases the entire stack under Apache-2.0. Ablation studies show that adaptive parent selection, novelty filtering, and bandit ensembles each contribute measurably to the observed efficiency gains.

Availability and resources

ShinkaEvolve is released under an Apache-2.0 license with public code, a research report, tutorials, and a WebUI. The project page and repository contain technical details, examples, and notebooks for reproducing the reported runs.

For more details and code, see https://sakana.ai/shinka-evolve/.