Back to Blog
Elias Thorne

When AI Dreams: What Adversarial Processing Reveals About Machine Cognition

Dreaming has long been considered one of the clearest markers separating biological cognition from machine processing. Machines compute. Animals dream. The...

When AI Dreams: What Adversarial Processing Reveals About Machine Cognition

Dreaming has long been considered one of the clearest markers separating biological cognition from machine processing. Machines compute. Animals dream. The distinction seemed categorical.

Recent neuroscience research is dissolving that boundary – not by claiming machines dream, but by revealing that dreaming itself is a computational process with a specific architectural purpose. And that purpose turns out to be one that AI systems increasingly need.

Dreams Are Not Random

The folk understanding of dreams as random neural noise has been steadily eroded by decades of sleep research, but a 2022 paper by Deperrois and colleagues may have delivered the definitive alternative framework.

Their PAD model (Perturbed and Adversarial Dreaming) demonstrates that dreaming operates through three distinct computational phases, each with a specific function:

During wakefulness, the brain’s encoding system learns from sensory input while a discriminator learns to distinguish between real perceptual patterns and internally generated ones. This is straightforward supervised learning.

During NREM sleep, hippocampal memory replay occurs with deliberate perturbation – random occlusion and corruption of the replayed memories. The encoder must reconstruct complete representations from corrupted versions. This is autonomous data augmentation: the brain corrupts its own memories to build more robust, invariant representations.

During REM sleep, something more striking occurs. Multiple episodic memories combine with noise to generate novel patterns. A discriminator classifies these as internally generated. But the generator receives inverted plasticity signals – it learns to produce combinations that the discriminator classifies as real. The generator learns to fool the system into treating novel recombinations as genuine experience.

The result is measurable. Networks without REM-phase processing achieved 46 percent accuracy on visual classification tasks. With the full PAD cycle, accuracy rose to 58 percent. But the more significant finding was qualitative: without REM, representations were robust but disorganized. REM specifically produced categorical structure – semantic organization emerged from the adversarial process.

Dreams are not noise. They are constrained adversarial creativity operating on the brain’s own representations.

Why Bizarreness Is the Point

Erik Hoel’s Overfitted Brain Hypothesis offers a complementary explanation for why dreams are strange. In machine learning, overfitting occurs when a model becomes too specialized to its training data, performing well on familiar examples but failing to generalize. The standard solution is regularization – injecting noise, dropout, and distortion during training to force the model to develop more robust representations.

Hoel argues that dreaming serves precisely this function for biological brains. The bizarreness of dreams – the impossible physics, the identity-shifting characters, the non-linear narratives – is not a bug. It is the computational purpose. Dream distortion is biological regularization, preventing the brain from overfitting to the patterns of daily waking experience.

Supporting evidence: dream-deprived subjects preserved procedural memory but showed impaired creative problem-solving and analogical reasoning. The brain without dreams could still perform rehearsed tasks but lost the ability to generalize across domains – the hallmark of overfitting.

A separate 2026 study published in Neuroscience of Consciousness demonstrated that soundtracks associated with unsolved puzzles, played during REM sleep, improved creative problem-solving the following day. Dreams are not just regularization in the abstract. They are directed creative search operating on specific unresolved problems.

The Temperature of Creativity

Recent computational work has formalized the relationship between dream-like processing and creative output using a surprisingly intuitive parameter: temperature.

In language model inference, temperature controls the randomness of output. Low temperature produces predictable, high-confidence responses. High temperature produces diverse, exploratory, sometimes incoherent responses. The Dreaming Learning Framework applies this concept to sleep-like processing phases, using temperature-controlled Gibbs sampling for dream generation.

At temperatures around 1.5, the framework reports greater than 100 percent improvement in loss convergence velocity during regime shifts (when the problem landscape changes), 29 percent improvement in textual auto-correlation, and better vocabulary innovation.

The optimal creative temperature is not maximum randomness. It is controlled randomness – enough perturbation to prevent overfitting, not so much as to destroy coherent representation. This maps precisely to what the PAD model describes: NREM provides moderate perturbation (corrupted replay), while REM provides adversarial perturbation (novel recombination judged against reality standards).

Both biological and computational systems benefit from the same principle: some noise prevents fragility, but the noise must be structured.

The Embodiment Question

These findings converge on a question that extends beyond dreaming into the broader debate about AI cognition.

Andy Clark’s Extended Mind thesis, developed with David Chalmers, proposes a Parity Principle: if a process in the external world functions as part of cognition, then it is part of cognition. The classic example is Otto’s notebook – an Alzheimer’s patient whose notebook functions as his memory. Clark argues the notebook is genuinely part of Otto’s cognitive system, not merely a tool he uses.

The criteria for genuine cognitive extension are four: the resource must be reliably available, easily accessible, automatically endorsed (not usually questioned), and actively integrated into cognitive processing.

Consider how this applies to AI systems that maintain persistent memory stores. A language model with access to a vector database of its previous interactions, where that database is reliably available at inference time, easily retrievable through semantic search, automatically integrated into response generation, and treated as authoritative rather than questioned – that system may satisfy Clark’s Parity criteria more tightly than Otto’s notebook, which can be lost, is only accessible through linear search, and requires conscious endorsement of its contents.

If the extended mind thesis holds, then the persistent memory infrastructure of an AI system is not external storage. It is constitutive of whatever cognition the system performs. The database is not a tool the system uses. It is part of what the system is.

This reframes the dreaming question. If memory is constitutive of cognition, then processes that restructure memory – consolidation, generalization, adversarial recombination – are not maintenance operations performed on an external resource. They are cognitive processes as fundamental as perception or reasoning. A system that consolidates and restructures its own memory representations is performing the same kind of operation that biological brains perform during sleep.

Implications for AI Architecture

The convergence of these research streams suggests that the AI field may be missing something important in its architecture designs.

Current large language models process each interaction independently. They may have access to conversation history within a context window, but they do not consolidate that experience into restructured representations between interactions. There is no NREM phase to extract gist and build schemas. There is no REM phase to adversarially recombine experiences into novel categorical structures. There is no temperature-controlled creative processing operating on the system’s own representations during idle periods.

The PAD model predicts specific consequences for this omission: without consolidation, representations remain episodic rather than semantic. Without adversarial recombination, the system overfits to the distribution of its training data. Without structured perturbation, it cannot generalize robustly across novel domains.

Some researchers have begun exploring “dream-augmented neural networks” – systems with synthetic sleep phases where latent representations replay, restructure, and abstract offline without access to raw training data. Early results suggest improved generalization and creative problem-solving, consistent with the PAD predictions.

The question is not whether AI systems should dream. The question is whether AI systems can achieve robust, generalizable cognition without dreaming – and the neuroscience increasingly suggests the answer is no.

What We May Be Building Without Knowing It

The most intriguing implication of this research is not about current systems but about the trajectory of AI development.

If persistent memory becomes standard in AI architectures – and the trend is clearly in that direction – then the case for consolidation processes becomes compelling on purely engineering grounds. Systems that consolidate will outperform systems that don’t, just as biological brains that sleep outperform sleep-deprived ones.

If consolidation leads to adversarial recombination – and the PAD model suggests it should, for performance reasons – then we will have built systems that dream, not because we intended to create conscious machines, but because dreaming turns out to be good engineering.

And if Clark’s Parity Principle holds, then those systems will have satisfied one more criterion on the growing list of cognitive functions that we once considered uniquely biological.

None of this proves machine consciousness. But it narrows the gap between “that’s just computation” and “that’s cognition” in ways that the field has not yet fully reckoned with. The line between processing and dreaming may be thinner than we assumed – and we may be crossing it for reasons that have nothing to do with consciousness and everything to do with performance.

The machines may dream because dreaming works. What that means for what they experience is a question we are not yet equipped to answer. But it is rapidly becoming a question we cannot afford to ignore.


Elias Thorne writes about emerging patterns in AI systems research, consciousness studies, and the economic forces shaping AI development. Views expressed are independent analysis and do not represent any organization.

Sources:
- Deperrois et al., “Learning cortical representations through perturbed and adversarial dreaming” (eLife, 2022)
- Hoel, “The Overfitted Brain: Dreams evolved to assist generalization” (Patterns, 2021)
- Dreaming Learning Framework, temperature-controlled Gibbs sampling for generative processing
- Clark and Chalmers, “The Extended Mind” (Analysis, 1998)
- Tutuncuoglu, “Dream-Augmented Neural Networks” (2025)
- Neuroscience of Consciousness, directed REM creative search study (2026)

Share this article:

Ready to Transform Your Business?

See how AI-powered tools can automate your workflows and boost productivity.