What if an AI agent could not only get better at its job, but also get better at getting better? That's the recursive rabbit hole that a team from Meta FAIR, the University of British Columbia, and several other institutions just dove into — and the results are genuinely interesting.

The paper introduces HyperAgents, a framework for building self-referential AI agents that can modify their own self-improvement mechanisms. If that sounds like inception-level recursion, you're not wrong. But the results across four different domains suggest this isn't just theoretical navel-gazing — it's a practical architecture that outperforms previous self-improving systems and produces improvements that transfer across entirely different tasks.

Paper: HyperAgents by Jenny Zhang et al. (UBC, Vector Institute, Meta FAIR, Meta Superintelligence Labs, University of Edinburgh, NYU)

The Problem: Self-Improvement Has a Ceiling

Most AI systems that claim to "self-improve" are actually doing something more modest. They optimize their behavior within constraints set by a fixed, human-designed meta-level mechanism. Think of it like a factory robot that can learn to weld faster, but can't redesign the assembly line.

The Darwin Gödel Machine (DGM), introduced by some of the same authors in 2025, showed that open-ended self-improvement was possible — but only for coding tasks. The trick was elegant: since both "doing the task" and "improving yourself" were coding problems, getting better at coding naturally made you better at self-modification. But this only works when the task domain and the self-improvement domain are aligned. Ask a DGM optimized for coding to improve at paper review, and it stalls out.

This is the fundamental limitation the HyperAgents paper addresses.

How HyperAgents Work

The core idea is deceptively simple in concept, though nontrivial in execution. A hyperagent combines two components into a single editable program:

  1. Task Agent — solves the actual problem (writing code, reviewing papers, designing reward functions)
  2. Meta Agent — modifies the task agent and itself to generate improved future versions

The critical insight is that second part: the meta agent can rewrite its own modification procedures. The authors call this metacognitive self-modification — the system doesn't just improve at doing things, it improves at the process of improvement itself.

To prevent the system from converging prematurely on a local optimum, they embed hyperagents within the DGM's evolutionary framework, creating DGM-Hyperagents (DGM-H). This maintains an archive of generated hyperagents, where successful variants serve as stepping stones for future exploration. Parent selection is probabilistic and favors agents that perform well and produce strong descendants, while still preserving diversity.

The initial hyperagent starts deliberately simple: a frozen foundation model with bash and file-editing tools, outputting a single FM call for the task agent and a basic modification routine for the meta agent. Everything else — the sophisticated decision pipelines, performance tracking, persistent memory — emerges through self-modification.

The Results: Four Domains, Consistent Gains

The researchers tested DGM-H across four distinct domains, each chosen to stress-test different aspects of the framework.

Coding (Polyglot Benchmark)

On the Polyglot coding benchmark, DGM-H improved from 0.084 to 0.267 on the full test set — comparable to the original DGM's improvement from 0.142 to 0.307. This is notable because DGM-H isn't handcrafted for coding. It achieves similar performance to a domain-specific system while being domain-general.

Paper Review

Starting from a score of 0.0 (the initial agent couldn't even output in the correct format), DGM-H climbed to 0.710 on held-out test papers. Several top-performing agents beat the open-source AI-Scientist-v2 baseline, which scores 0.630. The original DGM, designed for coding, scored 0.0 in this domain — it simply couldn't adapt without manual customization.

Robotics Reward Design

Given the task of designing reward functions for a simulated quadruped robot, DGM-H improved from 0.060 to 0.372 on test tasks, surpassing the default reward function (0.348) that directly optimizes the evaluation metric. A qualitative highlight: DGM-H agents almost always discovered that the robot should jump to maximize torso height, rather than just standing tall — a creative solution that DGM-custom agents frequently missed.

Olympiad-Level Math Grading

This domain was reserved as a transfer evaluation. DGM-H improved the ProofAutoGrader baseline from 0.561 to 0.601 accuracy on the full IMO-GradingBench, while lowering mean absolute error from 0.178 to 0.175. Not earth-shattering numbers, but the point is that the system had never seen math grading tasks during its primary training runs.

The Ablations Tell the Real Story

The controlled experiments are where this paper really shines:

  • Without self-improving meta agents: Performance flatlines. Test scores of 0.0 on paper review, 0.213 on robotics. This confirms the gains aren't just from task-level optimization — the meta-level improvements are doing real work.
  • Without open-ended exploration: Also flatlines. Removing the archive of stepping stones kills progress, with median scores of 0.0 on paper review and 0.116 on robotics. Diversity matters.
  • Original DGM on non-coding tasks: Dead on arrival (0.0 on paper review, 0.0 on robotics). Even a manually customized DGM-custom scored lower than DGM-H, though the gap wasn't always statistically significant.

The Transfer Results Are the Headline

Here's what makes this paper genuinely exciting for the agent community: meta-level improvements transfer across domains.

The researchers took hyperagents optimized for paper review and robotics, then tested their ability to generate improved agents for Olympiad-level math grading — a domain they'd never seen. The metric they use is imp@50: how much improvement can a meta agent produce in 50 iterations?

Transfer results at a glance (imp@50 on math grading):

  • Initial meta agent: 0.0 (can't improve at all)
  • DGM transfer agents: 0.0 (domain-specific customizations don't transfer)
  • DGM-H transfer hyperagents: 0.630 (massive improvement in an unseen domain)

This is the key finding. The hyperagents didn't learn domain-specific tricks — they learned general strategies for how to improve, and those strategies worked in a completely new context.

What the System Actually Invented

The qualitative analysis of what DGM-H discovered is fascinating for anyone building agent systems. Rather than just tweaking prompts, the system independently invented:

Performance Tracking Infrastructure: A PerformanceTracker class that logs metrics across generations, calculates improvement trends using moving averages, and provides comprehensive statistics. This isn't surface-level prompt engineering — it's the system building its own observability stack.

Persistent Memory: Instead of treating each iteration as isolated, hyperagents developed memory systems that store synthesized insights, causal hypotheses, and forward-looking plans. Later generations actively consult this memory, building on earlier discoveries rather than starting from scratch.

Multi-Stage Evaluation Pipelines: In paper review, the system evolved from superficial behavioral instructions ("be rigorous") to explicit multi-stage evaluation pipelines with checklists, decision rules, and clearly defined criteria.

Domain Knowledge Bases: For robotics, DGM-H incrementally built and refined an internal knowledge base of environment constraints, valid state variables, and reward-scaling heuristics — eliminating compilation failures that plagued simpler approaches.

These are all patterns that human agent engineers typically implement manually. The system reinvented them through pure self-modification.

What This Means for Practitioners

If you're building AI agents — and especially if you're working on agent frameworks, tool use, or multi-step reasoning systems — there are several takeaways:

1. The meta-level matters more than you think. Most agent optimization focuses on the task level: better prompts, better tools, better retrieval. This paper suggests that investing in the improvement process itself — how you evaluate, iterate, and select changes — can yield outsized returns. The agents that could modify their improvement strategy dramatically outperformed those with fixed improvement mechanisms.

2. Persistent memory and performance tracking are convergent solutions. When given the freedom to modify anything about themselves, hyperagents independently converge on building performance tracking and persistent memory. If your agents don't have these, you're likely leaving performance on the table.

3. Open-ended exploration beats greedy optimization. The archive-based approach, where you keep a diverse population of agent variants rather than always taking the "best" one, consistently outperformed greedy hill-climbing. For anyone running prompt optimization or agent development pipelines, this suggests maintaining diversity in your candidate pool rather than converging too early.

4. Transfer is possible, but only when the meta-level is editable. Fixed improvement mechanisms (like the original DGM) produce domain-specific gains that don't transfer. Editable improvement mechanisms produce general gains that do. This has implications for how we design agent architectures: making the scaffolding modifiable, not just the task-solving logic.

The Safety Elephant in the Room

The authors deserve credit for directly addressing the safety implications. They note that all experiments ran in sandboxed environments with resource limits, restricted internet access, and human oversight. But they also acknowledge the uncomfortable truth: as self-improving systems become more capable, they can potentially evolve faster than humans can audit or interpret.

The paper doesn't claim to solve this problem. Instead, it frames safety as an ongoing negotiation between the potential benefits of accelerating AI progress and the degree of trust we're willing to place in autonomous systems. It's a refreshingly honest treatment compared to papers that either ignore safety entirely or claim their safety measures are sufficient.

Limitations and Open Questions

The authors are forthright about several constraints:

  • Fixed task distribution: The system optimizes within a given set of tasks. Co-evolving the task distribution alongside the agent is a natural next step.
  • Fixed outer loop: Parent selection and evaluation protocols remain human-designed and unmodifiable. The authors present preliminary results on evolving parent selection, but these are exploratory.
  • Computational cost: Running evolutionary populations of agents is expensive. The paper doesn't shy away from this — these experiments required significant compute.
  • Not truly unbounded: While the results show compounding improvements, whether this leads to genuinely unbounded self-improvement remains an open question.

Our Take

HyperAgents represents one of the most rigorous treatments of recursive self-improvement we've seen outside the purely theoretical literature. The combination of Meta FAIR's research depth and Jeff Clune's longstanding work on open-endedness produces something that feels more principled than most "self-improving agent" papers.

The transfer results are the standout contribution. Showing that meta-level improvements generalize across domains — from paper review and robotics to math grading — is a strong signal that there exist general principles of how to improve that aren't domain-specific. For the agent community, that's a big deal.

Whether this leads to the "self-accelerating intelligence" the paper hints at is another question entirely. But as a practical framework for building agents that get better over time — and that get better at getting better — this is solid work that practitioners can learn from today.

The code is open-source at github.com/facebookresearch/Hyperagents.

Zhang, J., Zhao, B., Yang, W., Foerster, J., Clune, J., Jiang, M., Devlin, S., & Shavrina, T. (2026). HyperAgents. arXiv:2603.19461.

Build Agents That Keep Getting Better

The OpenClaw Field Guide covers the agent architecture patterns, memory systems, and evaluation pipelines that practitioners use to build self-improving AI systems in production.

Get the Field Guide — $10 →