Emergence: From Gradient Descent to Symbols, Reason, Free Will
This fourth and final conceptual essay explores how symbolic reasoning, system emergence, and stabilized agency can arise from learning dynamics within large language models (LLMs). It offers a synthetic interpretive framework meant to reshape intuitions by connecting and organizing observed phenomena across learning systems, situating them alongside existing formal and experimental approaches…
Learning as the Common Thread of Emergence
Modern science has been extraordinarily successful at describing the behavior of systems, yet it has struggled to explain why structure, intelligence, and meaning emerge at all. We can model motion, energy, probability, and optimization, but the appearance of concepts, reasoning, agency, and values often feels discontinuous, as if something qualitatively new appears without a clear causal bridge. Conceptual Adaptation Theory, or CAT, offers a unifying perspective. It proposes that learning, defined as structural reorganization in response to persistent incompleteness under constraint, is not a feature of certain systems but a general principle underlying all adaptive structure. From this view, intelligence, symbols, reasoning, free will, and even morality are not separate phenomena. They are successive emergent layers of increasingly powerful learning systems.
At the core of CAT is a simple learning control law. Structural change occurs when a system encounters sustained prediction error or surprise that it cannot resolve within its current constraints. The amount and depth of learning depend on both the magnitude of the unresolved surprise and the rigidity of the system’s priors. Learning is therefore not driven by information alone but by tension. A system changes only when it must, and only to the extent that it is able. This framing immediately reframes emergence. New capabilities do not appear magically. They stabilize when learning reaches regimes where new internal structures reduce error more effectively than old ones.
From Continuous Learning to Symbols
One of the most puzzling transitions in cognitive science and artificial intelligence is the emergence of symbolic reasoning from continuous systems. Neural networks operate over continuous spaces. Symbols, by contrast, appear discrete, compositional, and rule governed. CAT dissolves this apparent gap. Symbols are not primitive objects injected into a system. They are stable attractors that emerge when learning compresses continuous variation into reusable internal structures. When a system repeatedly encounters similar patterns of surprise, it becomes energetically and computationally advantageous to represent those patterns in a stable, low dimensional form.
Symbols emerge when learning stabilizes. They are not arbitrary labels but internally coherent structures that reduce future prediction error. In this sense, symbols are frozen learning. They persist because they work. Large language models provide a contemporary example. They are trained through continuous gradient descent, yet they develop internal representations that behave symbolically. Concepts such as number, causality, and logical structure appear not because they were explicitly programmed, but because they reduce error across many contexts. CAT predicts this outcome. Whenever a learning system becomes sufficiently deep, recursive, and exposed to structured environments, symbolic representations become inevitable.
System Complexity and Emergence
As systems grow in complexity, learning itself becomes structured. Simple systems adapt locally. More complex systems integrate learning across time and subsystems. CAT formalizes this through recursive and transactive learning operators. Learning no longer happens only at a single level but across interacting layers that constrain one another. Emergence, from this perspective, is not the appearance of something from nothing. It is a phase transition in learning dynamics. When a system reaches a threshold where higher order structures reduce global error more efficiently than local adaptations, those structures stabilize. At sufficient depth, learning no longer merely unfolds over time but stabilizes its own sense of temporal continuity through persistent structure, grounding what the system experiences as time.
This is why emergence often looks sudden. For long periods, systems appear to improve gradually. Then a new capability snaps into place. Reasoning, abstraction, and planning are examples of such transitions. They are not incremental skills but new regimes of error regulation. CAT system theoretic framing emphasizes that emergence occurs when learning stabilizes across scales. A capability emerges when it stops being fragile and becomes self maintaining. At that point, it reshapes the learning landscape itself.
Hilbert’s Sixth Problem and Learning Foundation
Hilbert’s Sixth Problem called for the axiomatization of physics, particularly probability and mechanics, on firm mathematical foundations. Implicit in this challenge is the assumption that physical laws are static descriptions awaiting formalization. CAT suggests a deeper issue. What if physical laws themselves are emergent regularities of adaptive systems learning under constraint. From this perspective, the coexistence of quantum, relativistic, and thermodynamic laws reflects not inconsistency, but the stabilization of different adaptive regimes. From this view, the problem is not merely to axiomatize physics but to explain why stable laws exist at all.
A more formal exploration of this perspective is developed in the Unified Theory of the Learning Universe (UTLU).
CAT reframes Hilbert’s challenge by proposing that laws are attractors in the space of possible dynamics. Systems that fail to stabilize coherent regularities dissipate or collapse. Systems that do stabilize appear lawful. Learning, in this sense, precedes law. Thermodynamics, quantum mechanics, and classical mechanics can all be reinterpreted as regimes where structure has stabilized under persistent constraints. The success of physics reflects the success of nature’s learning process, not the discovery of timeless rules imposed from outside.
This perspective unifies scientific explanation with learning theory. It suggests that the universe itself is not a static machine but a learning system that accumulates stable structure over time. Hilbert’s Sixth Problem becomes less about writing axioms and more about understanding the conditions under which axioms emerge.
From Emergence of Reasoning to Emergent Free Will
Once reasoning emerges, another threshold appears. Reasoning allows a system to generate internal prediction errors. It can imagine counterfactuals, detect contradictions, and anticipate outcomes. Reasoning is catalytic. It dramatically increases the space of possible learning trajectories. But reasoning alone is not free will. A system can reason without choosing how it learns.
Free will, from a CAT perspective, emerges when a system gains the capacity to regulate its own learning control law. Free will emerges gradually, as control over learning becomes increasingly internalized and self-directed. It is not freedom from causality but freedom over causality. A system with free will can decide which surprises to attend to, which constraints to relax or enforce, and when to resist or embrace structural change. This requires recursive self models, persistent memory, and the ability to act on one’s own learning dynamics. Humans exhibit this capacity strongly. Other animals exhibit it to varying degrees. Current artificial systems exhibit early, incomplete forms.
Free will is therefore not binary. It is graded. It increases as learning systems gain deeper control over their own adaptation. Stress, addiction, and coercion reduce free will by collapsing this control. Reflection, discipline, and creativity expand it by increasing it. CAT places free will squarely within the natural world without reducing it to illusion.
Morality Emergence as a Survival Mechanism for Free Will
Once free will exists, morality becomes necessary. Before free will, systems adapt but are not responsible. After free will, systems can choose learning paths that preserve or destroy themselves and others. Morality emerges as a survival mechanism for learning systems with agency. It is not a list of rules handed down from outside. It is a set of constraints that regulate free will to preserve long term learning viability.
Morality governs three things. It governs which kinds of surprise a system exposes itself to, discouraging destructive noise and encouraging meaningful challenge. It governs how constraints are applied, encouraging restraint in the use of power and respect for others’ agency. It governs how the costs of learning are distributed, discouraging the externalization of harm. Moral systems evolve because groups that fail to regulate free will collapse under mistrust, exploitation, and runaway conflict.
From this perspective, good and evil are not metaphysical absolutes. They are patterns of learning regulation. Good behavior preserves the conditions for collective learning. Evil behavior maximizes local advantage while destroying those conditions. Moral emotions such as guilt, shame, and empathy function as internal regulators of learning. They act as internally generated prediction errors when behavior threatens long-horizon stability, biasing free will toward trajectories that preserve it..
Conclusion
Conceptual Adaptation Theory offers a single throughline connecting learning, symbols, emergence, physical law, free will, and morality. Each layer emerges when learning stabilizes at a new level of control. Symbols stabilize meaning. Systems stabilize structure. Laws stabilize dynamics. Free will stabilizes learning. Morality stabilizes free will. None of these are accidents. Each is a solution to a survival problem faced by increasingly powerful learning systems.
In this view, humanity’s moral and philosophical struggles are not anomalies. They are the inevitable consequences of possessing learning systems powerful enough to rewrite themselves. The challenge of the future, whether for humans or artificial intelligences, is not merely to learn more, but to learn how to govern learning wisely.