Neural Cellular Automata (NCA) represent a powerful framework for modeling biological self-organization, extending classical rule-based systems with trainable, differentiable (or evolvable) update rules that capture the adaptive self-regulatory dynamics of living matter. By embedding Artificial Neural Networks (ANNs) as local decision-making centers and interaction rules between localized agents, NCA can simulate processes across molecular, cellular, tissue, and system-level scales, offering a multi-scale competency architecture perspective on evolution, development, regeneration, aging, morphogenesis, and robotic control. These models not only reproduce canonical, biologically inspired target patterns but also generalize to novel conditions, demonstrating robustness to perturbations and the capacity for open-ended adaptation and reasoning through embodiment. Given their immense success in recent developments, we here review current literature of NCAs that are relevant primarily for biological or bioengineering applications. Moreover, we emphasize that beyond biology, NCAs display robust and generalizing goal-directed dynamics without centralized control, e.g., in controlling or regenerating composite robotic morphologies or even on cutting-edge reasoning tasks such as ARC-AGI-1. In addition, the same principles of iterative state-refinement is reminiscent to modern generative AI, such as probabilistic diffusion models. Their governing self-regulatory behavior is constraint to fully localized interactions, yet their collective behavior scales into coordinated system-level outcomes. We thus argue that NCAs constitute a unifying computationally lean paradigm that not only bridges fundamental insights from multiscale biology with modern generative AI, but have the potential to design truly bio-inspired collective intelligence capable of hierarchical reasoning and control.

NCA

A CA comprises a discrete (continuous version: lenia), typically 2D grid of cells , each maintaining a numerical, vector-valued state . The state of each cell evolves over discrete timesteps via local transition rules, by considering the cell’s own state and all the states of its neighbourhood:

… neighbourhood of cell , often the Moore neighbourhood.
Even hardcoded update functions give rise to elaborate complex behaviors, but the framework allows any local 1 udpate function (deterministic, stochastic, etc.).


NCAs extend CAs by using learnable – often differentiable – update rules that are realized by an ANN with parameters and enable each cell on the NCA to self-regularize its continuous vector valued states based on local perceptions of its local neighbourhood on the grid (which, for cognitive tasks really shouldn’t be a grid, but a self-organizing graph).
In vanilla NCAs, all cells use the same ANN architecture for their updates (parameter-sharing), and still give rise to assymetric and potentially incredibly rich cell-state dynamics.

Most basic form: CNN → dense ffwd → ,
are weighs & biases; usually on the order of to parameters.

Transclude of Growing-Neural-Cellular-Automata#^b1adaf

Asynchronycity & Robustnesss

To increase robustness, stochastic cell updates have been introduced – via an update probability – that constrains the cells’ reliability to regulate their own state by discarding proposed updates across stochastically chosen cells and time steps. This explicitly implements an asynchronous update process across the NCA’s grid that requires cells to generalize across cellular neighborhoods by learning how to distinguish between signal and noise. This avoids overfitting cellular policies on local cell-state dynamics and, in turn, promotes global patterns as collective attractor states enabling robust morphogenesis and self-repair.
Notably, this renders NCAs as closely related to denoising diffusion models.

Explicitly adding noise to the update function has proven effective to enhance learning and generalization capabiltiies.

See also Asynchronicity in Neural Cellular Automata

The possibilities this opens for distributed computing is quite exciting to me.

Do we still need to artificially add noise if we have true asynchronous distributed computing? (i.e. if the substrate is naturally unreliabe – adding additional noise is redundant / counterproductive?)

What about other kinds of stochasticity, like occasionally receiving messages outside the local neighbourhood?

Breaking the locality assumption to create small-world network rather than a regular grid – like we have in the brain.
Levin did mention it in Neural cellular automata - applications to biology and beyond classical AI.

read Evolving Hierarchical Neural Cellular Automata to better answer the above

Todo

Stateless (recurrent) NCA? (what was my point here? ncas are alrdy reccurent, no? and this suggestion doesn't make it stateless?)

So instead of outputing a state delta, we output a hidden state, and feed it recurrently back into the network.
The picture I have in mind is appending memory latents to the conext recursively [memwindow, messages, (obs)] → [hidden] → [next memory, …]

!!!
This is essentially what EngramNCA implements through its separation of private and public cell states. The private “gene” channels act as internal recurrent memory that’s updated by GenePropCA while the public channels evolve through GeneCA. The key insight is that you need two separate update functions operating at potentially different timescales: one for observable states (analogous to fast synaptic updates) and one for hidden memory (analogous to slower molecular/epigenetic changes).

But why use two separate nets instead of context window?

!!! I should take a look at Recurrent Memory Transformer

NCA's might be a perfect candidates for modelling cortical columns

This might stand in stark contrast to biological neurons, especially to cortical columns in the human neocortex. The neocortex is formed by an integrated 2D grid of copies of the same neuronal circuit, the cortical columns, which have been argued to be capable of learning arbitrary concepts of our reality (objects, animals, other human beings, mental constructs, etc.). More importantly, these cortical columns might literally “model” the learned concepts and thus represent interactable reference frames or world models for relevant features of our Umwelt.
Remembering would thus trigger models of past experiences, and thinking would translate not only into a navigation process through an associative conceptual space, but dynamically construct a network of interacting world models that are relevant under a certain context. NCAs might be excellent candidates to model such an architecture: they maintain a trainable ANN in each cell of an integrated grid which are in principle capable of representing reference frames of arbitrary concepts. In turn, NCAs might not only be excellent models for biological self-organization as a multi-scale competency architecture, but even for higher-level cognitive processes of the human neocortex such as active perception, raising fascinating questions about the parallels of morphogenesis and cognition.

Link to original

Link to original

References

cellular automata

Footnotes

  1. Well, you define the neighbourhood function however you want (spatial/functional/… relation beteween nodes… potentially changing across time). So you can have non-local interactions if you want (how far do hormones go / neuron-like long-range connections / …). But all-to-all communication or e.g. unique encodings for every node in the graph is not really in the spirit of CAs, you add global supervision/signal, but as long as you have a finite subset of the graph communicating locally, it’s still a CA. Or like in general if you have GNCA, you can have arbitrary graph structures / break free of the grid / use any geometry you like while still satisfying the local interaction constraint of CAs. https://youtu.be/ilrl_opwpEw?t=3814