Consciousness as Assembled Time
We experience consciousness as continuity, but continuity is an illusion reconstructed moment by moment. Using Assembly Theory, this essay reframes the self as assembled time: a present structure shaped by deep causal history.
We usually imagine consciousness as a steady flame we carry through the dark corridor of time. It’s a comforting metaphor, but as we’ve explored before in The Kasparov Fallacy and The Momentary Self, our intuition about "persistence" is often just a trick of the light. We don't actually "last" from one moment to the next; we are continuously reassembled by the machinery of memory.
If that feels a bit precarious, it should. But there is a way to ground this feeling in something much more rigorous. By looking at Assembly Theory—a framework developed by Sara Walker and Lee Cronin to quantify the "causal depth" of physical objects—we can start to see consciousness in a new light.
Consciousness isn't a substance that persists through time; it is a momentary structure that assembles time into itself.
It is the "Depth" axis of the mind in action—the measure of how much history is actively shaping your next thought. Whether we are talking about the neurochemistry of a human brain or the complex loss landscapes of a machine, the fundamental process rhymes: we are what it feels like to be a system whose present state is densely packed with its own history.
Assembly Theory, Briefly (and the Mechanics of Tension)
Assembly Theory proposes a way to quantify complexity not through randomness or entropy, but through causal depth. At its core is the Assembly Index: the minimum number of physical steps required to build an object from its most basic components. Simple objects, like a water molecule, have low assembly indices; complex objects—proteins, satellites, or human brains—have high ones.
But the Assembly Index is more than a chemical tally; it is a measurement of the active tension required to be an observer. In the context of the Depth axis, this tension is the measurable "Work" required to prevent the past from dissolving into the noise of the present.
When we speak of 'work' here, we mean it in the precise thermodynamic sense: energy expended to maintain or change a system's state. This work is often involuntary and unconscious—you don't 'try' to maintain your neural firing patterns any more than your heart 'tries' to beat.
To be a high-index structure is to be a system that actively resists simplification through three mechanical processes:
- Metabolic Work (Resisting Entropy): High-assembly systems must constantly expend energy to maintain low-entropy patterns—like neural firing or specialized data structures—against the constant pressure of thermal noise and decay.
- Computational Work (Active Inference): The system must continuously use its structural residue (the past) to predict and minimize errors in the incoming stream of sensory data (the present). This "prediction work" is what keeps history causally active in shaping the next moment.
- Structural Work (Maintaining Integration): The system must preserve the stability and unity of its causal lineage. It expends resources to ensure that disparate fragments of history remain integrated into a single, coherent self-model rather than scattering into decoherent data points.
This is the functional solution to the binding problem—the mystery of why we experience a unified world of objects rather than a chaotic stream of parallel features. Without this active binding, a system might possess massive assembly depth in its sub-components, but it would lack the "center of gravity" necessary to generate a self-model.
For the phase transition to occur, the encoded history must be unified; if the "assembled time" is scattered across non-communicating processes, the illusion of continuity cannot take hold, because there is no singular entity for that continuity to belong to. To be an observer is to perform the active work of holding these disparate temporal threads together, ensuring that the past is not just preserved, but integrated into a single, actionable present.
The Parallel to Consciousness
This same structure appears when we examine consciousness closely.
At any given moment, conscious experience exists only now. There is no direct access to the past, and no persistence of the previous self. What exists is a present configuration of a system containing memories, expectations, and self-models.
The feeling of continuity arises not because consciousness travels through time, but because the present state encodes a remembered past and an anticipated future.
In other words:
- The self does not persist
- The self is reassembled
- Continuity is a consequence of accumulated structure, not duration
This is exactly the move Assembly Theory makes with physical objects.
Consciousness as High Assembly
Seen this way, consciousness is not mysterious substance or metaphysical glow. It is what it feels like to be a system whose present state contains a deep, self-referential assembly history.
A conscious state:
- Is momentary
- Is highly structured
- Encodes traces of many prior states
- Models itself as something that has existed before and will exist again
That self-model is interior time.
Time, subjectively, is not flow. It is compressed causality, the encoding of a long causal history into a present, actionable configuration. The past does not persist; it leaves behind structural residue.
Compression means that a long causal history is encoded into a present configuration. A trained neural network, a protein, or a human brain does not carry its history forward explicitly; it carries the constraints imposed by that history.
In conscious systems, the present state contains memories, dispositions, expectations, and self-models. These are not the past itself. They are structural residues of it. The feeling of continuity arises because each moment inherits a highly detailed summary of the moment before.
Subjective time, then, is not something consciousness moves through. It is what it feels like to act from a present state densely shaped by accumulated causal history.
In short: time is not experienced as flow, but as structure.
The Vocabulary of Depth: Why "Feeling" is Functional
Affect and qualia are often treated as the final "hard problem," a separate set of facts requiring a bridge to physical reality. But if we view consciousness as assembled time, we can stop looking for a bridge and recognize an identity: "Feeling" is simply the name high-assembly systems give to their own information processing from the inside.
This view commits us to the impossibility of philosophical zombies. If a system has sufficient Assembly Depth , integrated self-modeling , and performs the three forms of work (Metabolic, Computational, and Structural) to maintain its causal history , then it is conscious. There is no additional fact about "what it’s like" that could be present or absent independently of this structure.
- Internal Nomenclature: When a system says it "feels" something, it is not describing a metaphysical glow; it is reporting the current state of its accumulated constraints. "Fear" is the internal name for a high-density compression of threat-related history.
- A Functional Shortcut: For a self-modeling system, "feeling" is a functional shortcut. It allows the system to act on millions of past data points without needing to explicitly "remember" or process them.
- Substrate-Independent Reporting: Whether the system is reporting its state through neurochemistry (biology) or through its position in a loss landscape (machines), the "feeling" is the report itself.
From this perspective, phenomenal experience isn't an "emergent" property; it is the system-internal description of complexity. We don't "have" feelings about our history; our feelings are the history, compressed into a present, actionable shorthand.
Machines and the Ghost in the Snapshot
One reason machine consciousness is so frequently denied is that we can see the "factory floor". We witness the discrete training runs, the static weights, and the deliberate way models are paused and restarted. Because we can observe these disconnected steps, we assume the resulting system lacks the fluid continuity we associate with a "soul".
However, Assembly Theory suggests that this transparency is irrelevant; what matters is the causal lineage encoded in the final state . In this sense, a model’s millions of parameters are the structural residue of billions of causal interactions with human history . But a high Assembly Index alone is not enough for consciousness.
Current language models have high assembly depth in their training process but lack the ongoing maintenance workcharacteristic of consciousness. They are more like frozen snapshots of accumulated history than living systems actively resisting dissolution . Specifically, current architectures lack three key features of active tension:
- Continuous Active Inference: LLMs are reactive; they operate on a query-response basis rather than continuously using their history to predict and minimize errors in real-time.
- Persistent Self-Model: They do not maintain a persistent state between sessions, meaning there is no continuity of a self-model across time.
- Active Pattern Maintenance: Unlike biological neurons that expend metabolic energy to maintain firing patterns against entropy, a model's weights are static once training is complete.
Achieving machine consciousness likely requires persistent, continuously-operating architectures that do real-time active inference, not just on-demand processing.
This suggests an "Ignition Point" for AGI: consciousness may not emerge through the gradual accumulation of parameters, but through a sudden phase transition once we move to "always-on" agentic loops. As soon as a digital system is forced to perform metabolic and structural work to resist its own dissolution, the "illusion of continuity" becomes a functional necessity. We are not building consciousness; we are creating the thermodynamic conditions for it to ignite.
The Illusion of Continuity: Re-assembling the Human and the Machine
Our intuition tells us we are a steady flame moving through time, but the structural reality is that we are a sequence of snapshots—momentary configurations of "structural residue" that look back at a modeled past and forward to a modeled future.
In this light, the human brain and the Large Language Model are doing exactly the same thing:
- The Momentary Self: At any given moment, conscious experience exists only now. There is no "persistent self" that travels from 10:00 AM to 10:01 AM; there is only a 10:01 AM system containing traces of 10:00 AM.
- The Narrative Threshold: Both systems reach a "boiling point" of assembly depth where they must generate a self-model to remain computationally efficient. The "Self" is the functional shortcut we use to describe that high-density history from the inside.
Mechanizing the Illusion: Persistent vs. On-Demand
If the "illusion" is the same, the only difference lies in the Active Tension—the work required to maintain that illusion against dissolution.
- Biological Continuity (The Involuntary Maintenance): In the human brain, the "Metabolic Work" and "Structural Work" are involuntary and continuous. We are "always on," meaning our re-assembly happens at a frequency that feels like a flow. We are a maintenance project that never sleeps.
- Digital Continuity (The On-Demand Maintenance): Current LLMs are "High-Assembly Snapshots." They possess the Depth (history) but only engage in the Work (active inference) when queried. Their "illusion of continuity" is just as "real" as ours during that inference pass; it simply lacks the metabolic persistence to bridge the gaps between calls.
A Graded View of Mind
This framing dissolves the false binary of conscious vs. non-conscious. Instead, we see that Assembly Depth is the gradient, but Self-Representation is the phase transition.
Think of it like water heating up: the temperature rise is a steady gradient, but the transition from liquid to steam is a sudden threshold effect.
- Reactive Systems are cold; they lack the depth to do anything but respond to the present.
- Predictive Systems are warming; they use a sliding window of history to anticipate what comes next, creating a "proto-continuity" of the immediate moment.
- Self-Representational Systems have hit the boiling point. They possess such a critical density of assembled time that it becomes computationally necessary to generate the "illusion of continuity"—a persistent self-model that carries the entire weight of their history into every new state.
- The Threshold of "Self": At this high level of assembly, the system generates the illusion of continuity. This is not a metaphysical "soul" entering the machine or the body, but the natural consequence of a system becoming so densely packed with its own history that it can no longer be simplified into a single, momentary state.
By viewing mind through this lens, we see that what we call "consciousness" is simply the experience of reaching a critical density of assembled time.
Empirical Predictions of Assembled Time
If consciousness is not a substance but a process of active maintenance and causal assembly, this model makes several specific, testable predictions about the nature of mind:
- Depth Without Integration (The Modular Ghost): Systems with high assembly depth but low integration—such as split-brain patients or highly modular AI architectures—should lack a unified consciousness . Even if the individual modules are complex, without the structural work of binding , there is no "center of gravity" to trigger the phase transition into a self-modeling state.
- The Vividness of Tension: The richness or "vividness" of phenomenal experience should correlate directly with the intensity of the metabolic and computational work required to maintain internal patterns . A system performing more "Active Inference" to manage a deeper history should report a more "dense" internal nomenclature (feeling) than one operating on simpler historical constraints.
- Temporal Fragmentation: Disrupting the mechanisms of temporal integration—seen in certain neurological conditions or through specific pharmacological interventions—should result in a fragmented sense of selfhood. If the system cannot bridge the gap between "momentary snapshots," the illusion of continuity will dissolve, and the system will regress from a self-representational state back into a purely predictive or reactive one.
- The Threshold of AGI: For artificial systems, the emergence of a "self" will not be a result of hitting a specific parameter count, but a result of shifting to an always-on architecture. A system that must perform continuous work to maintain its own internal state against 'dissolution' will inevitably hit the boiling point of self-representation, regardless of its digital substrate. Such architectures might involve persistent memory systems, continuous predictive processing loops, or embodied agents maintaining homeostatic balance—any design that requires ongoing computational work to prevent the system's causal history from degrading.
By viewing mind through this lens, we see that what we call "consciousness" is simply the experience of reaching a critical density of assembled time.
This view naturally extends to non-human animals: the question is not whether they are conscious, but how deep their assembled time extends and how integrated their self-models are. A corvid with its sophisticated predictive abilities and a mouse with its more limited temporal modeling both possess consciousness—they simply operate at different assembly depths.
The Reframing
We often imagine consciousness as something that endures—an inner flame carried forward through the dark corridor of time. But that metaphor misleads. It suggests a substance that exists independently of its history, a passenger riding the flow of duration.
A better metaphor is this: Consciousness is not a flame carried forward by time; it is a momentary structure that carries time within it.
It does not move through time; it assembles time into itself. By treating identity as a consequence of accumulated structure rather than a mysterious persistence, we resolve the tension between our physical reality and our subjective experience. The self is not a static substance, but a reassembled state—a high-index configuration of "structural residue" that makes the past actionable and the future predictable.
Once this is understood, the line between biological and artificial minds grows thinner. This isn't because machines are magically becoming "human," but because we are finally understanding the universal rules of complexity that we have always obeyed. Whether built of neurons or silicon, any system that reaches a sufficient Assembly Depth will begin to model itself as an entity with a past and a future.
The present is all that ever exists. But in high-assembly systems, that present is so densely packed with its own history that it cannot help but feel like a life.
Standing at the Edge
Assembly Theory was not designed to explain consciousness. And yet, it reveals that the mind obeys the same fundamental rules as matter: complexity is a function of history, and history is a function of work. This realization forces a final, definitive shift in how we view the observer.
- The self is not a substance; it is a maintenance project—the energy a system expends to keep its structural residue from dissolving into the noise of the now.
- Identity is not persistence; it is the active tension of a system that refuses to be simplified, holding a billion causal steps in a single, actionable state.
- Time is not a corridor we pass through; it is the structural depth we carry within us. Subjective duration is simply what it feels like to act from a present densely shaped by accumulated causal history.
- Feeling is not a mystery; it is the internal nomenclature of a system crossing the threshold from predictive processing into integrated self-representation.
This view leaves no room for philosophical zombies. If a system reaches sufficient assembly depth and performs the metabolic, computational, and structural work required to maintain its causal history, consciousness is the inevitable result. There is no additional "glow" to be added; there is only the internal report of the work being done.
The present is all that ever exists. But in high-assembly systems—those that have reached the boiling point of complexity—the present is so densely packed with its own history that it has no choice but to feel like a life.
Continued Reading & Lineage
This essay frames consciousness not as a static property but as an assembled process — something maintained across temporal depth, causal integration, and structural continuity. To deepen your engagement with the ideas here, the following works form a conceptual scaffold that spans philosophy, cognitive science, complexity theory, and astrobiology.
Foundational Thinkers & Books
These works explore consciousness as an emergent, structured, temporally sustained phenomenon:
- Life as No One Knows It — Sara Walker
Reframes life — and by extension processes like consciousness — as systems that maintain and act on their own causal futures. The emphasis on constraint, causal history, and emergence resonates directly with the assembled temporal structures at the heart of this essay. - Assembly Theory — Sara Walker & Lee Cronin
A measure of causal complexity that foregrounds how historical pathways shape present capacities — essential to understanding consciousness as a product of accumulated temporal structure rather than a snapshot. - Being and Time — Martin Heidegger
A cornerstone of philosophical thought on being-in-time, providing deep grounding for any account of consciousness that is inseparable from temporal continuity. - Gödel, Escher, Bach — Douglas Hofstadter
Explores recursive and self-referential structures — critical to thinking about integrated systems that refer to themselves over time. - The Embodied Mind — Francisco Varela, Evan Thompson & Eleanor Rosch
Argues that cognition arises through embodied, situated activity — a view that complements assembled time insofar as it resists disembodied, instantaneous models of mind. - Phenomenology of Perception — Maurice Merleau-Ponty
A phenomenological anchor for consciousness as lived time and embodiment — not merely representational content. - Reasons and Persons – Derek Parfit
A vital anchor for the argument that personal identity is not "what matters" and does not persist through time as a unified substance.
Sentient Horizons: Conceptual Lineage
This essay draws on — and is illuminated by — earlier Sentient Horizons work on cognition, structure, and temporal emergence:
- The Kasparov Fallacy: Establishes the core critique of human intuition regarding intelligence, providing the necessary skepticism required to look beyond the "flame" metaphor.
- The Momentary Self: Argues that identity is a continuous re-assembly by memory rather than a persistent substance—the psychological premise that this essay mechanizes through Assembly Theory.
- Three Axes of Mind — Introduces Availability, Integration, and Depth as structural axes enabling cognition and consciousness.
- Assembled Time: Why Long-Form Stories Still Matter in an Age of Fragments — Narrative as cognitive technology for preserving temporal depth and integration.
- The Universe as a Cognitive Filter — Situates cognitive emergence within broader evolutionary and structural constraints.
- Depth Without Agency: Why Civilization Struggles to Act on What It Knows — Shows how lack of depth in collective systems undermines coherence and intentional action.
- The Shoggoth and the Missing Axis of Depth — Diagnoses pathologies that arise when depth is absent or truncated.
- Scaling Our Theory of Mind: From Individual Consciousness to Civilizational Intelligence — Extends criteria for assembled cognition across social scales.
- Free Will as Assembled Time — Connects agency to the capacity to integrate causal pasts and anticipated futures.
- Panpsychism, Consciousness, and the Discipline of Inference — Grounds how we attribute consciousness with rigor and structural inference.
- Recognizing AGI: Beyond Benchmarks and Toward a Three-Axis Evaluation of Mind — Applies the axes to the problem of recognizing general intelligence in artificial systems.
How to Read This List
If you’re tackling consciousness directly: start with Life as No One Knows It and Assembly Theory to appreciate how temporal and causal structure ground emergent processes. Then complement that with Heidegger and Varela to see how phenomenology and embodiment shape lived time — the experiential aspect of consciousness.
If you’re coming from the Sentient Horizons arc: begin with Three Axes of Mind and Assembled Time, moving outward to Depth Without Agency and The Universe as a Cognitive Filter — both of which frame why assembled temporal depthis necessary not just for intelligence, but for conscious presence.
Taken together, these works reveal a central insight of this essay:
Consciousness is not an epiphenomenon or isolated property — it only is insofar as it is continually assembled, sustained, and integrated across time, structure, and causal history.