year: 2025/05
paper: https://arxiv.org/abs/2505.22954
website: https://sakana.ai/dgm/ || sakana-ai-dgm
code: https://github.com/jennyzzt/dgm
connections: Gödel Machine, SSA, quality diversity, open-endedness, Jenny Zhang, Jeff Clune


Takes the Gödel Machine’s idea (a self-referential, self-improving AI) but drops the impractical requirement of formal proof search.
Instead: empirical evaluation (like SSA) + population-based open-ended search.

A coding agent that modifies its own Python codebase to become a better coding agent. Self-improvement is the downstream task: better coding ability directly means better self-modification ability. The LLM is the mutation operator (“primitive learning algorithm”), instead of blind parameter tweaks, modifications are proposed by a foundation model that reads the codebase and writes patches.

The archive is the key difference from SSA-style single-system self-improvement. It lets the search traverse fitness valleys: the path to the best SWE-bench agent went through ancestors that performed worse than their parents. A greedy approach (always branch from the best) would never have found them.