Tiny Recursive Model: How a 7M-Parameter Solver Outperforms Much Larger LLMs on ARC-AGI

What TRM is and why it matters

Samsung SAIT Montreal introduced the Tiny Recursive Model (TRM), a compact recursive reasoner with roughly 7 million parameters that challenges larger autoregressive LLMs on symbolic reasoning benchmarks. TRM is an iterative draft–revise solver that maintains a latent scratchpad and a current solution embedding, repeatedly refining its candidate answers through recursion rather than autoregressive token decoding.

Core architectural changes

TRM departs from the prior Hierarchical Reasoning Model (HRM) by simplifying the design into a single tiny recurrent core. Instead of HRM’s two-module hierarchy and fixed-point gradient approximation, TRM uses a 2-layer network that jointly updates a latent scratchpad z and a current solution embedding y. The model alternates between a “think” update and an “act” update:

This think→act block is unrolled up to 16 times with deep supervision and a learned halting head used during training, while tests use the full unroll. Signals are propagated across steps through the paired state (y, z).

TRM also moves away from HRM’s one-step implicit fixed-point gradient approximation by backpropagating through the entire recursion during training, which the authors report as essential for generalization.

Architectures and training details

Benchmark performance

TRM shows surprising performance gains on several benchmarks compared to larger models and prior specialized architectures:

These are trained-from-scratch direct-prediction models on small but heavily augmented datasets, not few-shot prompting. ARC remains the canonical target, with broader leaderboard context and thresholds tracked by the ARC Prize Foundation.

Why a 7M model can beat larger LLMs on these tasks

Key takeaways

For more technical detail, see the paper on arXiv: https://arxiv.org/pdf/2510.04871v1