levinlab.dev/Machine or Living?

Machine or Living?

This tool analyzes discussions against questions about intelligence, sentience, and cognition, based on "Discussions of machine vs. living intelligence need more clarity" by Nicolas Rouleau and Michel Levin.

Technological Approach to Mind Everywhere: An Experimentally-Grounded Framework for Understanding Diverse Bodies and Minds

Synthetic biology and bioengineering provide the opportunity to create novel embodied cognitive systems (otherwise known as minds) in a very wide variety of chimeric architectures combining evolved and designed material and software. These advances are disrupting familiar concepts in the philosophy of mind, and require new ways of thinking about and comparing truly diverse intelligences, whose composition and origin are not like any of the available natural model species. In this Perspective, I introduce TAME—Technological Approach to Mind Everywhere—a framework for understanding and manipulating cognition in unconventional substrates. TAME formalizes a non-binary (continuous), empirically-based approach to strongly embodied agency. TAME provides a natural way to think about animal sentience as an instance of collective intelligence of cell groups, arising from dynamics that manifest in similar ways in numerous other substrates. When applied to regenerating/developmental systems, TAME suggests a perspective on morphogenesis as an example of basal cognition. The deep symmetry between problem-solving in anatomical, physiological, transcriptional, and 3D (traditional behavioral) spaces drives specific hypotheses by which cognitive capacities can increase during evolution. An important medium exploited by evolution for joining active subunits into greater agents is developmental bioelectricity, implemented by pre-neural use of ion channels and gap junctions to scale up cell-level feedback loops into anatomical homeostasis. This architecture of multi-scale competency of biological systems has important implications for plasticity of bodies and minds, greatly potentiating evolvability. Considering classical and recent data from the perspectives of computational science, evolutionary biology, and basal cognition, reveals a rich research program with many implications for cognitive science, evolutionary biology, regenerative medicine, and artificial intelligence. Keywords: regeneration, basal cognition, bioelectricity, gap junctions, synthetic morphology, bioengineering Glossary The following definitions of terms used in this paper (in alphabetical order) represent ways of thinking about specific terminology in the context of the proposed TAME framework. These terms have many definitions in other frameworks and are tightly interwoven, and it is likely impossible to do them full justice at this point in time (or provide uncontroversial definitions that everyone will agree capture everything of importance). Moreover, much like a theorem and its component statements, the utility of these highly-related concepts is maximized by the entire set taken together, not by crisp demarcations of any one term. The below definitions are not claimed to be uniquely correct, but merely useful; this field is still sufficiently young with respect to very basic questions, which excessively sharp definitions can limit more than they enable. •Agency—a set of properties closely related to decision-making and adaptive action which determine the degree to which optimal ways to relate to the system (in terms of communication, prediction, and control) require progressively higher-level models specified in terms of scale of goals, stresses, capabilities, and preferences of that System as an embodied Self acting in various problem spaces. This view of agency is related to those of autopoiesis (Maturana and Varela, 1980) and anticipatory systems (Rosen, 1985). •Consciousness—the first-person phenomenal experience of any Self—that which makes my toothache irreducibly different to me than anyone else’s toothache or third-person descriptions of toothaches. The degree and content of consciousness is “what it is like” to be that Self, as opposed to studying it from the outside, whether or not the Self is advanced enough to be able to verbalize it or to think about it (Nagel, 1974). Consciousness here is not meant to necessarily indicate advanced, reflexive, verbal self-consciousness but rather the basal sentience (sense-process-respond loop) which is taken to be a continuum. Moreover, because all cognitive agents are inevitably made of parts, we are all collective intelligences in a strong sense (Schwitzgebel, 2015)—what it is like to be you is exactly what it’s like to be a (particularly organized) collection of cells. •Cognition—all of the activities undertaken by a Self, at whatever scale and of whatever material implementation, that underlie its gathering, processing, and acting on information for the purposes of adaptive action and perdurance against dissipation. Components include active inference, learning, and basal goal-directed activity, as well as complex cognitive skills such as symbolic reasoning, composition of concepts, language, and meta-cognition. •Decision—an event during the traversal of some relevant space by a system’s state which is efficiently modeled as a choice between diverse options. The degree of “decision-making” of any given system is proportional to the spatio-temporal and complexity distance between the events that eventually gave rise to a specific outcome and the outcome itself. Advanced Selves have inputs to their decision-making machinery that are counterfactual future states. The scale at which one defines appropriate inputs (stimuli) to a system is whatever scale is most efficient for understanding the resulting decisions (Noble, 2012; Pezzulo and Levin, 2016; Flack, 2017). •Mind—the functional, dynamic aspect of a Self that results from all of its cognitive and somatic activities, which represents the propensities for certain types of actions and possesses some degree of sentience as a first-person perspective that perdures across changes in the material components of the body. •Intelligence—the functional ability to solve problems in various spaces (not necessarily in 3D space), not tied to specific implementations, anatomical structures, or time scales. The degree of intelligence (IQ) is proportional to competency in navigating these spaces, including especially the ability to identify paths that temporarily lead further from the goal state but eventually enable better results. Advanced intelligence exploits additional levels of self-modeling which enables multiple levels of virtual modeling of the Self and its outside world (counterfactual thought), anxiety, and creativity (identifying opportunities, as opposed to only solving problems existing right now). In particular, by focusing on the functional aspects of intelligence, and by recognizing that there is no intelligent agent that is not made of parts, Collective Intelligence is generalized here (emphasizing the architecture of functional connections between subunits) and is not viewed as a radically distinct natural kind. •Maslow’s Hierarchy of Needs–a motivational theory of psychology that focuses on the relative types of preferences and goals which human (or other) systems pursue at various stages and scales of observation (Maslow, 1943). It also stresses degrees of integration and the modulation of higher levels by the level of stress in subunits. •Self—a coherent system emerging within a set of integrated parts that serves as the functional owner of associations, memories, and preferences, and acts to accomplish goals in specific problem spaces where those goals belong to the collective and not to any individual sub-component. Selves are defined by the spatio-temporal scale and nature of the types of goals they can pursue—their “cognitive light cone.” They have functional boundaries and material implementations but are not identical with any specific type of substrate, and can overlap within other Selves at the same, higher, and lower-level Selves. A Self is a theoretical construct posited by external systems (such as scientists, engineers, and conspecifics) and by systems themselves (via internal self-models), which facilitates prediction and adaptive behavior by serving as an efficient, high-level target for intervention and control strategies. •Stress—a system-level state which serves as a driver for homeostatic loops (operating over a variable that is progressively reduced as activity gets the system closer to its desired region of action space). The spatio-temporal and complexity scale of events that can possibly stress a system are a good indicator of that system’s cognitive sophistication. Stress can arise via discord between external states and the Self’s needs, between sensory stimuli and expectations, or between the goals of multiple subsystems within an agent, either within or across levels of organization. Thus, geometric frustration (Sadoc and Mosseri, 2007) and material scientists’ notions of stress as a high-level determinant of system behavior over time (Batterman and Rice, 2014; Batterman, 2015) are minimal examples of the fundamental concept of Stress, on the same continuum as metabolic stress in bacteria, competing cellular alignment forces in planar polarity of tissues, and “true psychological stress” in organisms. Introduction All known cognitive agents are collective intelligences, because we are all made of parts; biological agents in particular are not just structurally modular, but made of parts that are themselves agents in important ways. There is no truly monadic, indivisible yet cognitive being: all known minds reside in physical systems composed of components of various complexity and active behavior. However, as human adults, our primary experience is that of a centralized, coherent Self which controls events in a top-down manner. That is also how we formulate models of learning (“the rat learned X”), moral responsibility, decision-making, and valence: at the center is a subject which has agency, serves as the locus of rewards and punishments, possesses (as a single functional unit) memories, exhibits preferences, and takes actions. And yet, under the hood, we find collections of cells which follow low-level rules via distributed, parallel functionality and give rise to emergent system-level dynamics. Much as single celled organisms transitioned to multicellularity during evolution, the single cells of an embryo construct de novo, and then operate, a unified Self during a single agent’s lifetime. The compound agent supports memories, goals, and cognition that belongs to that Self and not to any of the parts alone. Thus, one of the most profound and far-reaching questions is that of scaling and unification: how do the activities of competent, lower-level agents give rise to a multiscale holobiont that is truly more than the sum of its parts? And, given the myriad of ways that parts can be assembled and relate to each other, is it possible to define ways in which truly diverse intelligences can be recognized, compared, and understood? Here, I develop a framework to drive new theory and experiment in biology, cognition, evolution, and biotechnology from a multi-scale perspective on the nature and scaling of the cognitive Self. An important part of this research program is the need to encompass beings beyond the familiar conventional, evolved, static model animals with brains. The gaps in existing frameworks, and thus opportunities for fundamental advances, are revealed by a focus on plasticity of existing forms, and the functional diversity enabled by chimeric bioengineering. To illustrate how this framework can be applied to unconventional substrates, I explore a deep symmetry between behavior and morphogenesis, deriving hypotheses for dynamics that up- and down-scale Selves within developmental and phylogenetic timeframes, and at the same time strongly impact the speed of the evolutionary process itself (Dukas, 1998). I attempt to show how anatomical homeostasis can be viewed as the result of the behavior of the swarm intelligence of cells, and provides a rich example of how an inclusive, forward-looking technological framework can connect philosophical questions with specific empirical research programs. The philosophical context for the following perspective is summarized in Table 1 (see also Glossary), and links tightly to the field of basal cognition (Birch et al., 2020) via a fundamentally gradualist approach. It should be noted that the specific proposals for biological mechanisms that scale functional capacity are synergistic with, but not linearly dependent on, this conceptual basis. The hypotheses about how bioelectric networks scale cell computation into anatomical homeostasis, and the evolutionary dynamics of multi-scale competency, can be explored without accepting the “minds everywhere” commitments of the framework. However, together they form a coherent lens onto the life sciences which helps generate testable new hypotheses and integrate data from several subfields. TABLE 1. The core tenets of TAME. •Continuum of cognitive capacities—no binary categories, no bright line separating true cognition from “just physics,” as is clear from evolutionary process and ability to bioengineer chimeras between any two “natural kinds.” •Mature frameworks must apply to truly diverse intelligences—beyond the examples from Earth’s phylogenetic tree based on brains, we must be able to consider and compare agents across the option space of designed and evolved combinations of living, non-living, and software components at all scales. •Selves exist across a continuum of persuadability, and it is an empirical question as to where on this axis any given system lies (revealed by the ratio of prediction and control vs. effort and knowledge that needs to be input, for any given way of relating to that system). •Selves are not fixed, permanent agents—their substrate can remodel radically during their lifetime; the owner of memories and preferences, and the subject that interprets rewards and punishments, is malleable and plastic. •The core of being a Self is the ability to pursue goals. Selves can be nested and overlapping, cooperating and competing both laterally and across levels. Each higher-level self deforms the option space for the lower level Selves, enabling them to follow energy minimization to achieve outcomes that look inevitable and simple at one scale, while serving intelligent goals at a higher scale. •Intelligence is the degree of competency of navigating any space (not just the familiar 3D space of motility), including morphospace, transcriptional space, physiological space, etc., toward desirable regions, while avoiding being trapped in local minima. Estimates of intelligence of any system are observer-dependent, and say as much about the observer and their limitations as they do about the system itself. Open in a new tab For the purposes of this paper, “cognition” refers not only to complex, self-reflexive advanced cognition or metacognition, but is used in the less conservative sense that recognizes many diverse capacities for learning from experience (Ginsburg and Jablonka, 2021), adaptive responsiveness, self-direction, decision-making in light of preferences, problem-solving, active probing of their environment, and action at different levels of sophistication in conventional (evolved) life forms as well as bioengineered ones (Rosenblueth et al., 1943; Lyon, 2006; Bayne et al., 2019; Levin et al., 2021; Lyon et al., 2021; Figure 1). For our purposes, cognition refers to the functional computations that take place between perception and action, which allow the agent to span a wider range of time (via memory and predictive capacity, however much it may have) than its immediate now, which enable it to generalize and infer patterns from instances of stimuli—precursors to more advanced forms of recombining concepts, language, and logic. FIGURE 1. FIGURE 1 Open in a new tab Diverse, multiscale intelligence. (A) Biology is organized in a multi-scale, nested architecture of molecular pathways. (B) These are not merely structural, but also computational: each level of this holarchy contains subsystems which exhibit some degree of problem-solving (i.e., intelligent) activity, on a continuum such as the one proposed by Rosenblueth et al. (1943). (C) At each layer of a given biosystem, novel components can be introduced of either biological or engineered origin, resulting in chimeric forms that have novel bodies and novel cognitive systems distinct from the typical model species on the Earth’s phylogenetic lineage. Images in panels (A,C) by Jeremy Guay of Peregrine Creative. Image in panel (B) was created after Rosenblueth et al. (1943). The framework, TAME—Technological Approach to Mind Everywhere—adopts a practical, constructive engineering perspective on the optimal place for a given system on the continuum of cognitive sophistication. This gives rise to an axis of persuadability (Figure 2), which is closely related to the Intentional Stance (Dennett, 1987) but made more explicit in terms of functional engineering approaches needed to implement prediction and control in practice. Persuadability refers to the type of conceptual and practical tools that are optimal to rationally modify a given system’s behavior. The origin story (designed vs. evolved), composition, and other aspects are not definitive guides to the correct level of agency for a living or non-living system. Instead, one must perform experiments to see which kind of intervention strategy provides the most efficient prediction and control (thus, one aim should be generalizing the human-focused Turing Test and other IQ metrics into a broader agency detection toolkit, which perhaps could itself be implemented by a useful algorithm). FIGURE 2. FIGURE 2 Open in a new tab The axis of persuadability. A proposed way to visualize a continuum of agency, which frames the problem in a way that is testable and drives empirical progress, is via an “axis of persuadability”: to what level of control (ranging from brute force micromanagement to persuasion by rational argument) is any given system amenable, given the sophistication of its cognitive apparatus? Here are shown only a few representative waypoints. On the far left are the simplest physical systems, e.g., mechanical clocks (A). These cannot be persuaded, argued with, or even rewarded/punished—only physical hardware-level “rewiring” is possible if one wants to change their behavior. On the far right (D) are human beings (and perhaps others to be discovered) whose behavior can be radically changed by a communication that encodes a rational argument that changes the motivation, planning, values, and commitment of the agent receiving this. Between these extremes lies a rich panoply of intermediate agents, such as simple homeostatic circuits (B) which have setpoints encoding goal states, and more complex systems such as animals which can be controlled by signals, stimuli, training, etc., (C). They can have some degree of plasticity, memory (change of future behavior caused by past events), various types of simple or complex learning, anticipation/prediction, etc. Modern “machines” are increasingly occupying right-ward positions on this continuum (Bongard and Levin, 2021). Some may have preferences, which avails the experimenter of the technique of rewards and punishments—a more sophisticated control method than rewiring, but not as sophisticated as persuasion (the latter requires the system to be a logical agent, able to comprehend and be moved by arguments, not merely triggered by signals). Examples of transitions include turning the sensors of state outward, to include others’ stress as part of one’s action policies, and eventually the meta-goal of committing to enhance one’s agency, intelligence, or compassion (increase the scope of goals one can pursue). A more negative example is becoming sophisticated enough to be susceptible to a “thought that breaks the thinker” (e.g., existential or skeptical arguments that can make one depressed or even suicidal, Gödel paradoxes, etc.)—massive changes can be made in those systems by a very low-energy signal because it is treated as information in the context of a complex host computational machinery. These agents exhibit a degree of multi-scale plasticity that enables informational input to make strong changes in the structure of the cognitive system itself. The positive flip side of this vulnerability is that it avails those kinds of minds with a long term version of free will: the ability through practice and repeated effort to change their own thinking patterns, responses to stimuli, and functional cognition. This continuum is not meant to be a linear scala naturae that aligns with any kind of “direction” of evolutionary progress—evolution is free to move in any direction in this option space of cognitive capacity; instead, this scheme provides a way to formalize (for a pragmatic, engineering approach) the major transitions in cognitive capacity that can be exploited for increased insight and control. The goal of the scientist is to find the optimal position for a given system. Too far to the right, and one ends up attributing hopes and dreams to thermostats or simple AIs in a way that does not advance prediction and control. Too far to the left, and one loses the benefits of top-down control in favor of intractable micromanagement. Note also that this forms a continuum with respect to how much knowledge one has to have about the system’s details in order to manipulate its function: for systems in class A, one has to know a lot about their workings to modify them. For class B, one has to know how to read-write the setpoint information, but does not need to know anything about how the system will implement those goals. For class C, one doesn’t have to know how the system modifies its goal encodings in light of experience, because the system does all of this on its own—one only has to provide rewards and punishments. Images by Jeremy Guay of Peregrine Creative. Our capacity to find new ways to understand and manipulate complex systems is strongly related to how we categorize agency in our world. Newton didn’t invent two terms—gravity (for terrestrial objects falling) and perhaps shmavity (for the moon)—because it would have lost out on the much more powerful unification. TAME proposes a conceptual unification that would facilitate porting of tools across disciplines and model systems. We should avoid quotes around mental terms because there is no absolute, binary distinction between it knows and it “knows”—only a difference in the degree to which a model will be useful that incorporates such components. Given this perspective, below I develop hypotheses about invariants that unify otherwise disparate-seeming problems, such as morphogenesis, behavior, and physiological allostasis. I take goals (in the cybernetic sense) and stressors (as a system-level result of distance from one’s goals) as key invariants which allow us to study and compare agents in truly diverse embodiments. The processes which scale goals and stressors form a positive feedback loop with modularity, thus both arising from, and potentiating the power of, evolution. These hypotheses suggest a specific way to understand the scaling of cognitive capacity through evolution, make interesting predictions, and suggest novel experimental work. They also provide ways to think about the impending expansion of the “space of possible bodies and minds” via the efforts of bioengineers, which is sure to disrupt categories and conclusions that have been formed in the context of today’s natural biosphere. What of consciousness? It is likely impossible to understand sentience without understanding cognition, and the emphasis of this paper is on testable, empirical impacts of ways to understand cognition in all of its guises. By enabling the definition, detection, and comparison of cognition and intelligence, in diverse substrates beyond standard animals, we can enhance the range of embodiments in which sentience may result. In order to move the field forward via empirical progress, the focus of most of the discussion below is on ways to think about cognitive function, not on phenomenal or access consciousness [in the sense of the “Hard Problem” (Chalmers, 2013)]. However, I return to this issue at the end, discussing TAME’s view of sentience as fundamentally tied to goal-directed activity, only some aspects of which can be studied via third person approaches. The main goal is to help advance and delineate an exciting emerging field at the intersection of biology, philosophy, and the information sciences. By proposing a new framework and examining it in a broad context of now physically realizable (not merely logically possible) living structures, it may be possible to bring conceptual, philosophical thought up to date with recent advances in science and technology. At stake are current knowledge gaps in evolutionary, developmental, and cell biology, a new roadmap for regenerative medicine, lessons that could be ported to artificial intelligence and robotics, and broader implications for ethics. Cognition: Changing the Subject Even advanced animals are really collective intelligences (Couzin, 2007, 2009; Valentini et al., 2018), exploiting still poorly-understood scaling and binding features of metazoan architectures that share a continuum with looser swarms that have been termed “liquid brains” (Sole et al., 2019). Studies of “centralized control” focus on a brain, which is in effect a network of cells performing functions that many cell types, including bacteria, can do (Koshland, 1983). The embodied nature of cognition means that the minds of Selves are dependent on a highly plastic material substrate which changes not only on evolutionary time scales but also during the lifetime of the agent itself. The central consequence of the composite nature of all intelligences is that the Self is subject to significant change in real-time (Figure 3). This means both slow maturation through experience (a kind of “software” change that doesn’t disrupt traditional ways of thinking about agency), as well as radical changes of the material in which a given mind is implemented (Levin, 2020). The owner, or subject of memories, preferences, and in more advanced cases, credit and blame, is very malleable. At the same time, fascinating mechanisms somehow ensure the persistence of Self (such as complex memories) despite drastic alterations of substrate. For example, the massive remodeling of the caterpillar brain, followed by the morphogenesis of an entirely different brain suitable for the moth or beetle, does not wipe all the memories of the larva but somehow maps them onto behavioral capacities in the post-metamorphosis host, despite its entirely different body (Alloway, 1972; Tully et al., 1994; Sheiman and Tiras, 1996; Armstrong et al., 1998; Ray, 1999; Blackiston et al., 2008). Not only that, but memories can apparently persist following the complete regeneration of brains in some organisms (McConnell et al., 1959; Corning, 1966; Shomrat and Levin, 2013) such as planaria, in which prior knowledge and behavioral tendencies are somehow transferred onto a newly-constructed brain. Even in vertebrates, such as fish (Versteeg et al., 2021) and mammals (von der Ohe et al., 2006), brain size and structure can change repeatedly during their lifespan. This is crucial to understanding agency and intelligence at multiple scales and in unfamiliar embodiments because observations like this begin to break down the notion of Selves as monadic, immutable objects with a privileged scale. Becoming comfortable with biological cognitive agents that are malleable in terms of form and function (change radically during the lifetime of an individual) makes it easier to understand the origins and changes of cognition during evolution or as the result of bioengineering effort. FIGURE 3. FIGURE 3 Open in a new tab Cognitive Selves can change in real-time. (A) Caterpillars metamorphose into butterflies, going through a process in which their body, brain, and cognitive systems are drastically remodeled during the lifetime of a single agent. Importantly, memories remain and persist through this process (Blackiston et al., 2015). (B) Planaria cut into pieces regenerate, with each piece re-growing and remodeling precisely what is needed to form an entire animal. (C) Planarians derived from tail fragments of trained worms still retain original information, illustrating the ability of memories to move across tissues and be reimprinted on newly-developing brains (Corning, 1966, 1967; Shomrat and Levin, 2013). Images by Jeremy Guay of Peregrine Creative. This little-studied intersection between regeneration/remodeling and cognition highlights the fascinating plasticity of the body, brain, and mind; traditional model systems in which cognition is mapped onto a stable, discrete, mature brain are insufficient to fully understand the relationship between the Self and its material substrate. Many scientists study the behavioral properties of caterpillars, and of butterflies, but the transition zone in-between, from the perspective of philosophy of mind and cognitive science, provides an important opportunity to study the mind-body relationship by changing the body during the lifetime of the agent (not just during evolution). Note that continuity of being across drastic biological remodeling is not only relevant for unusual cases in the animal kingdom, but is a fundamental property of most life—even humans change from a collection of cells to a functional individual, via a gradual morphogenetic process that constructs an active Self in real time. This has not been addressed in biology, and likewise not yet in computer science, where machine learning approaches use static neural networks (there is not a formalism for altering artificial neural networks’ architecture on the fly). What are the invariants that enable a Self to persist (and be recognizable by third-person investigations) despite such change? Memory is a good candidate (Shoemaker, 1959; Ameriks, 1976; Figure 3). However, at least certain kinds of memories can be transferred between individuals, by transplants of brain tissue or molecular engrams (Pietsch and Schneider, 1969; McConnell and Shelby, 1970; Bisping et al., 1971; Chen et al., 2014; Bedecarrats et al., 2018; Abraham et al., 2019). Importantly, the movement of memories across individual animals is only a special case of the movement of memory in biological tissue in general. Even when housed in the same “body,” memories must move between tissues—for example, in a trained planarian’s tail fragment re-imprinting its learned information onto the newly regenerated brain, or the movement of memories onto new brain tissue during metamorphosis. In addition to the spatial movement and re-mapping of memories onto new substrates, there is also a temporal component, as each memory is really an instance of communication between past and future Selves. The plasticity of biological bodies, made of cells that die, are born, and significantly rearrange their tissue architecture, suggests that the understanding of cognition is fundamentally a problem of collective intelligence: to understand how stable cognitive structures can persist and map onto swarm dynamics, with preferences and stressors that scale from those of their components. This is applicable even to such a “stable” form as the human brain, which is often spoken of as a single Subject of experience and thought. First, the gulf between planarian regeneration/insect metamorphosis and human brains is going to be bridged by emerging therapeutics. It is inevitable that stem cell therapies for degenerative brain diseases (Forraz et al., 2013; Rosser and Svendsen, 2014; Tanna and Sachan, 2014) will confront us with humans whose brains are partially replaced by the naïve progeny of cells that were not present during the formation of memories and personality traits in the patient. Even prior to these advances, it was clear that phenomena such as dissociative identity disorder (Miller and Triggiano, 1992), communication with non-verbal brain hemispheres in commissurotomy patients (Nagel, 1971; Montgomery, 2003), conjoined twins with fused brains (Gazzaniga, 1970; Barilan, 2003), etc., place human cognition onto a continuous spectrum with respect to the plasticity of integrated Selves that reside within a particular biological tissue implementation. Importantly, animal model systems are now providing the ability to harness that plasticity for functional investigations of the body-mind relationship. For example, it is now easy to radically modify bodies in a time-scale that is much faster than evolutionary change, to study the inherent plasticity of minds without eons of selection to shape them to fit specific body architectures. When tadpoles are created to have eyes on their tails, instead of their heads, they are still readily able to perform visual learning tasks (Blackiston and Levin, 2013; Blackiston et al., 2017). Planaria can readily be made with two (or more) brains in the same body (Morgan, 1904; Oviedo et al., 2010), and human patients are now routinely augmented with novel inputs [such as sensory substitution (Bach-y-Rita et al., 1969; Bach-y-Rita, 1981; Danilov and Tyler, 2005; Ptito et al., 2005)] or novel effectors, such as instrumentized interfaces allowing thought to control engineered devices such as wheelchairs in addition to the default muscle-driven peripherals of their own bodies (Green and Kalaska, 2011; Chamola et al., 2020; Belwafi et al., 2021). The central phenomenon here is plasticity: minds are not tightly bound to one specific underlying architecture (as most of our software is today), but readily mold to changes of genomic defaults. The logical extension of this progress is a focus on self-modifying living beings and the creation of new agents in which the mind:body system is simplified by entirely replacing one side of the equation with an engineered construct. The benefit would be that at least one half of the system is now well-understood. For example, in hybrots, animal brains are functionally connected to robotics instead of their normal body (Reger et al., 2000; Potter et al., 2003; Tsuda et al., 2009; Ando and Kanzaki, 2020). It doesn’t even have to be an entire brain—a plate of neurons can learn to fly a flight simulator, and it lives in a new virtual world (DeMarse and Dockendorf, 2005; Manicka and Harvey, 2008; Beer, 2014), as seen from the development of closed-loop neurobiological platforms (Demarse et al., 2001; Potter et al., 2005; Bakkum et al., 2007b; Chao et al., 2008; Rolston et al., 2009a,b). These kinds of results are reminiscent of Philosophy 101’s “brain in a vat” experiment (Harman, 1973). Brains adjust to driving robots and other devices as easily as they adjust to controlling a typical, or highly altered, living body because minds are somehow adapted and prepared to deal with body alterations—throughout development, metamorphosis and regeneration, and evolutionary change. The massive plasticity of bodies, brains, and minds means that a mature cognitive science cannot just concern itself with understanding standard “model animals” as they exist right now. The typical “subject,” such as a rat or fruit fly, which remains constant during the course of one’s studies and is conveniently abstracted as a singular Self or intelligence, obscures the bigger picture. The future of this field must expand to frameworks that can handle all of the possible minds across an immense option space of bodies. Advances in bioengineering and artificial intelligence suggest that we or our descendants will be living in a world in which Darwin’s “endless forms most beautiful” (this Earth’s N = 1 ecosystem outputs) are just a tiny sample of the true variety of possible beings. Biobots, hybrots, cyborgs, synthetic and chimeric animals, genetically and cellularly bioengineered living forms, humans instrumentized to knowledge platforms, devices, and each other—these technologies are going to generate beings whose body architectures are nothing like our familiar phylogeny. They will be a functional mix of evolved and designed components; at all levels, smart materials, software-level systems, and living tissue will be integrated into novel beings which function in their own exotic Umwelt. Importantly, the information that is used to specify such beings’ form and function is no longer only genetic—it is truly “epigenetic” because it comes not only from the creature’s own genome but also from human and non-human agents’ minds (and eventually, robotic machine-learning-driven platforms) that use cell-level bioengineering to generate novel bodies from genetically wild-type cells. In these cases, the genetics are no guide to the outcome (which highlights some of the profound reasons that genetics is hard to use to truly predict cognitive form and function even in traditional living species). Now is the time to begin to develop ways of thinking about truly novel bodies and minds, because the technology is advancing more rapidly than philosophical progress. Many of the standard philosophical puzzles concerning brain hemisphere transplants, moving memories, replacing body/brain parts, etc. are now eminently doable in practice, while the theory of how to interpret the results lags. We now have the opportunity to begin to develop conceptual approaches to (1) understand beings without convenient evolutionary back-stories as explanations for their cognitive capacities (whose minds are created de novo, and not shaped by long selection pressures toward specific capabilities), and (2) develop ways to analyze novel Selves that are not amenable to simple comparisons with related beings, not informed by their phylogenetic position relative to known standard species, and not predictable from an analysis of their genetics. The implications range across insights into evolutionary developmental biology, advancing bioengineering and artificial life research, new roadmaps for regenerative medicine, ability to recognize exobiological life, and the development of ethics for relating to novel beings whose composition offers no familiar phylogenetic touchstone. Thus, here I propose the beginnings of a framework designed to drive empirical research and conceptual/philosophical analysis that will be broadly applicable to minds regardless of their origin story or internal architecture. Technological Approach to Mind Everywhere: A Proposal for a Framework The Technological Approach to Mind Everywhere (TAME) framework seeks to establish a way to recognize, study, and compare truly diverse intelligences in the space of possible agents. The goal of this project is to identify deep invariants between cognitive systems of very different types of agents, and abstract away from inessential features such as composition or origin, which were sufficient heuristics with which to recognize agency in prior decades but will surely be insufficient in the future (Bongard and Levin, 2021). To flesh out this approach, I first make explicit some of its philosophical foundations, and then discuss specific conceptual tools that have been developed to begin the task of understanding embodied cognition in the space of mind-as-it-can-be (a sister concept to Langton’s motto for the artificial life community—“life as it can be”) (Langton, 1995). Philosophical Foundations of an Approach to Diverse Intelligences One key pillar of this research program is the commitment to gradualism with respect to almost all important cognition-related properties: advanced minds are in important ways generated in a continuous manner from much more humble proto-cognitive systems. On this view, it is hopeless to look for a clear bright line that demarcates “true” cognition (such as that of humans, great apes, etc.) from metaphorical “as if cognition” or “just physics.” Taking evolutionary biology seriously means that there is a continuous series of forms that connect any cognitive system with much more humble ones. While phylogenetic history already refutes views of a magical arrival of “true cognition” in one generation, from parents that didn’t have it (instead stretching the process of cognitive expansion over long time scales and slow modification), recent advances in biotechnology make this completely implausible. For any putative difference between a creature that is proposed to have true preferences, memories, and plans and one that supposedly has none, we can now construct in-between, hybrid forms which then make it impossible to say whether the resulting being is an Agent or not. Many pseudo-problems evaporate when a binary view of cognition is dissolved by an appreciation of the plasticity and interoperability of living material at all scales of organization. A definitive discussion of the engineering of preferences and goal-directedness, in terms of hierarchy requirements and upper-directedness, is given in McShea (2013, 2016). For example, one view is that only biological, evolved forms have intrinsic motivation, while software AI agents are only faking it via functional performance [but don’t actually care (Oudeyer and Kaplan, 2007, 2013; Lyon and Kuchling, 2021)]. But which biological systems really care—fish? Single cells? Do mitochondria (which used to be independent organisms) have true preferences about their own or their host cells’ physiological states? The lack of consensus on this question in classical (natural) biological systems, and the absence of convincing criteria that can be used to sort all possible agents to one or the other side of a sharp line, highlight the futility of truly binary categories. Moreover, we can now readily construct hybrid systems that consist of any percentage of robotics tightly coupled to on-board living cells and tissues, which function together as one integrated being. How many living cells does a robot need to contain before the living system’s “true” cognition bleeds over into the whole? On the continuum between human brains (with electrodes and a machine learning converter chip) that drive assistive devices (e.g., 95% human, 5% robotics), and robots with on-board cultured human brain cells instrumentized to assist with performance (5% human, 95% robotics), where can one draw the line—given that any desired percent combination is possible to make? No quantitative answer is sufficient to push a system “over the line” because there is no such line (at least, no convincing line has been proposed). Interesting aspects of agency or cognition are rarely if ever Boolean values. Instead of a binary dichotomy, which leads to impassable philosophical roadblocks, we envision a continuum of advancement and diversity in information-processing capacity. Progressively more complex capabilities [such as unlimited associative learning, counterfactual modeling, symbol manipulation, etc., (Ginsburg and Jablonka, 2021)] ramp up, but are nevertheless part of a continuous process that is not devoid of proto-cognitive capacity before complex brains appear. Specifically, while major differences in cognitive function of course exist among diverse intelligences, transitions between them have not been shown to be binary or rapid relative to the timescale of individual agents. There is no plausible reason to think that evolution produces parents that don’t have “true cognition” but give rise to offspring that suddenly do, or that development starts with an embryo that has no “true preferences” and sharply transitions into an animal that does, etc. Moreover, bioengineering and chimerization can produce a smooth series of transitional forms between any two forms that are proposed to have, or not have, any cognitive property. Thus, agents gradually shift (during their lifetime, as result of development, metamorphosis, or interactions with other agents, or during evolutionary timescales) between great transitions in cognitive capacity, expressing and experiencing intermediate states of cognitive capacity that must be recognized by empirical approaches to study them. A focus on the plasticity of the embodiments of mind strongly suggests this kind of gradualist view, which has been expounded in the context of evolutionary forces controlling individuality (Godfrey-Smith, 2009; Queller and Strassmann, 2009). Here the additional focus is on events taking place within the lifetime of individuals and driven by information and control dynamics. The TAME framework pushes experimenters to ask “how much” and “what kind of” cognition any given system might manifest if we interacted with it in the right way, at the right scale of observation. And of course, the degree of cognition is not a single parameter that gives rise to a scala naturae but a shorthand for the shape and size of its cognitive capacities in a rich space (discussed below). The second pillar of TAME is that there is no privileged material substrate for Selves. Alongside familiar materials such as brains made of neurons, the field of basal cognition (Nicolis et al., 2011; Reid et al., 2012, 2013; Beekman and Latty, 2015; Baluška and Levin, 2016; Boussard et al., 2019; Dexter et al., 2019; Gershman et al., 2021; Levin et al., 2021; Lyon et al., 2021) has been identifying novel kinds of intelligences in single cells, plants, animal tissues, and swarms. The fields of active matter, intelligent materials, swarm robotics, machine learning, and someday, exobiology, suggest that we cannot rely on a familiar signature of “big vertebrate brain” as a necessary condition for mind. Molecular phylogeny shows that the specific components of brains pre-date the evolution of neurons per se, and life has been solving problems long before brains came onto the scene (Buznikov et al., 2005; Levin et al., 2006; Jekely et al., 2015; Liebeskind et al., 2015; Moran et al., 2015). Powerful unification and generalization of concepts from cognitive science and other fields can be achieved if we develop tools to characterize and relate to a wide diversity of minds in unconventional material implementations (Damasio, 2010; Damasio and Carvalho, 2013; Cook et al., 2014; Ford, 2017; Man and Damasio, 2019; Baluska et al., 2021; Reber and Baluska, 2021). Closely related to that is the de-throning of natural evolution as the only acceptable origin story for a true Agent [many have proposed a distinction between evolved living forms vs. the somehow inadequate machines which were merely designed by man (Bongard and Levin, 2021)]. First, synthetic evolutionary processes are now being used in the lab to create “machines” and modify life (Kriegman et al., 2020a; Blackiston et al., 2021). Second, the whole process of evolution, basically a hill-climbing search algorithm, results in a set of frozen accidents and meandering selection among random tweaks to the micro-level hardware of cells, with impossible to predict large-scale consequences for the emergent system level structure and function. If this short-sighted process, constrained by many forces that have nothing to do with favoring complex cognition, can give rise to true minds, then so can a rational engineering approach. There is nothing magical about evolution (driven by randomizing processes) as a forge for cognition; surely we can eventually do at least as well, and likely much better, using rational construction principles and an even wider range of materials. The third foundational aspect of TAME is that the correct answer to how much agency a system has cannot be settled by philosophy—it is an empirical question. The goal is to produce a framework that drives experimental research programs, not only philosophical debate about what should or should not be possible as a matter of definition. To this end, the productive way to think about this a variant of Dennett’s Intentional Stance (Dennett, 1987; Mar et al., 2007), which frames properties such as cognition as observer-dependent, empirically testable, and defined by how much benefit their recognition offers to science (Figure 2). Thus, the correct level of agency with which to treat any system must be determined by experiments that reveal which kind of model and strategy provides the most efficient predictive and control capability over the system. In this engineering (understand, modify, build)-centered view, the optimal position of a system on the spectrum of agency is determined empirically, based on which kind of model affords the most efficient way of prediction and control. Such estimates are, by their empirical nature, subject to revision by future experimental data and conceptual frameworks, and are observer-dependent (not absolute). A standard methodology in science is to avoid attributing agency to a given system unless absolutely necessary. The mainstream view (e.g., Morgan’s Canon) is that it’s too easy to fall into a trap of “anthropomorphizing” systems with only apparent cognitive powers, when one should only be looking for models focused on mechanistic, lower levels of description that eschew any kind of teleology or mental capacity (Morgan, 1903; Epstein, 1984). However, analysis shows that this view provides no useful parsimony (Cartmill, 2017). The rich history of debates on reductionism and mechanism needs to be complemented with an empirical, engineering approach that is not inappropriately slanted in one direction on this continuum. Teleophobia leads to Type 2 errors with respect to attribution of cognition that carry a huge opportunity cost for not only practical outcomes like regenerative medicine (Pezzulo and Levin, 2015) and engineering, but also ethics. Humans (and many other animals) readily attribute agency to systems in their environment; scientists should be comfortable with testing out a theory of mind regarding various complex systems for the exact same reason—it can often greatly enhance prediction and control, by recognizing the true features of the systems with which we interact. This perspective implies that there is no such thing as “anthropomorphizing” because human beings have no unique essential property which can be inappropriately attributed to agents that have none of it. Aside from the very rare trivial cases (misattributing human-level cognition to simpler systems), we must be careful to avoid the pervasive, implicit remnants of a human-centered pre-scientific worldview in which modern, standard humans are assumed to have some sort of irreducible quality that cannot be present in degrees in slightly (or greatly) different physical implementations (from early hominids to cyborgs etc.). Instead, we should seek ways to naturalize human capacities as elaborations of more fundamental principles that are widely present in complex systems, in very different types and degrees, and to identify the correct level for any given system. Of course, this is just one stance, emphasizing experimental, not philosophical, approaches that avoid defining impassable absolute differences that are not explainable by any known binary transition in body structure or function. Others can certainly drive empirical work focused specifically on what kind of human-level capacities do and do not exist in detectable quantity in other agents. Avoiding philosophical wrangling over privileged levels of explanation (Ellis, 2008; Ellis et al., 2012; Noble, 2012), TAME takes an empirical approach to attributing agency, which increases the toolkit of ways to relate to complex systems, and also works to reduce profligate attributions of mental qualities. We do not say that a thermos knows whether to keep something hot or cold, because no model of thermos cognition does better than basic thermodynamics to explain its behavior or build better thermoses. At the same time, we know we cannot simply use Newton’s laws to predict the motion of a (living) mouse at the top of a hill, requiring us to construct models of navigation and goal-directed activity for the controller of the mouse’s behavior over time (Jennings, 1906). Under-estimating the capacity of a system for plasticity, learning, having preferences, representation, and intelligent problem-solving greatly reduces the toolkit of techniques we can use to understand and control its behavior. Consider the task of getting a pigeon to correctly distinguish videos of dance vs. those of martial arts. If one approaches the system bottom-up, one has to implement ways to interface to individual neurons in the animal’s brain to read the visual input, distinguish the videos correctly, and then control other neurons to force the behavior of walking up to a button and pressing it. This may someday be possible, but not in our lifetimes. In contrast, one can simply train the pigeon (Qadri and Cook, 2017). Humanity has been training animals for millennia, without knowing anything about what is in their heads or how brains work. This highly efficient trick works because we correctly identified them as learning agents, which allows us to offload a lot of the computational complexity of any task onto the living system itself, without micromanaging its components. What other systems might this remarkably powerful strategy apply to? For example, gene regulatory networks (GRNs) are a paradigmatic example of “genetic mechanism,” often assumed to be tractable only by hardware (requiring gene therapy approaches to alter promoter sequences that control network connectivity, or adding/removing gene nodes). However, being open to the possibility that GRNs might actually be on a different place on this continuum suggests an experiment in which they are trained for new behaviors with specific combinations of stimuli (experiences). Indeed, recent analyses of biological GRN models reveal that they exhibit associative and several other kinds of learning capacity, as well as pattern completion and generalization (Watson et al., 2010, 2014; Szilagyi et al., 2020; Biswas et al., 2021). This is an example in which an empirical approach to the correct level of agency for even simple systems not usually thought of as cognitive suggests new hypotheses which in turn open a path to new practical applications (biomedical strategies using associative regimes of drug pulsing to exploit memory and address pharmacoresistance by abrogating habituation, etc.). We next consider specific aspects of the framework, before diving into specific examples in which it drives novel empirical work. Specific Conceptual Components of the Technological Approach to Mind Everywhere Framework A useful framework in this emerging field should not only serve as a lens with which to view data and concepts (Manicka and Levin, 2019b), but also should drive research in several ways. It needs to first specify definitions for key terms such as a Self. These are not meant to be exclusively correct—the definitions can co-exist with others, but should identify a claim as to what is an essential invariant for Selves (and what other aspects can diverge), and how it intersects with experiment. The fundamental symmetry unifying all possible Selves should also facilitate direct comparison or even classification of truly diverse intelligences, sketching the markers of Selfhood and the topology of the option space within which possible agents exist. The framework should also help scientists derive testable claims about how borders of a given Self are determined, and how it interacts with the outside world (and other agents). Finally, the framework should provide actionable, semi-quantitative definitions that have strong implications and constrain theories about how Selves arise and change. All of this must facilitate experimental approaches to determine the empirical utility of this approach. The TAME framework takes the following as the basic hallmarks of being a Self: the ability to pursue goals, to own compound (e.g., associative) memories, and to serve as the locus for credit assignment (be rewarded or punished), where all of these are at a scale larger than possible for any of its components alone. Given the gradualist nature of the framework, the key question for any agent is “how well,” “how much,” and “what kind” of capacity it has for each of those key aspects, which in turn allows agents to be directly compared in an option space. TAME emphasizes defining a higher scale at which the (possibly competent) activity of component parts gives rise to an emergent system. Like a valid mathematical theorem which has a unique structure and existence over and above any of its individual statements, a Self can own, for example, associative memories (that bind into new mental content experiences that occurred separately to its individual parts), be the subject of reward or punishment for complex states (as a consequence of highly diverse actions that its parts have taken), and be stressed by states of affairs (deviations from goals or setpoints) that are not definable at the level of any of its parts (which of course may have their own distinct types of stresses and goals). These are practical aspects that suggest ways to recognize, create, and modify Selves. Selves can be classified and compared with respect to the scale of goals they can pursue [Figure 4, described in detail in Levin (2019)]. In this context, the goal-directed perspective adopted here builds on the work of Rosenblueth et al. (1943); Nagel (1979); and Mayr (1992), emphasizing plasticity (ability to reach a goal state from different starting points) and persistence (capacity to reach a goal (Schlosser, 1998) state despite perturbations). FIGURE 4. FIGURE 4 Open in a new tab Unconventional goal-directed agents and the scaling of the cognitive Self. (A) The minimal component of agency is homeostasis, for example the ability of a cell to execute the Test-Operate-Exit (Pezzulo and Levin, 2016) loop: a cycle of comparison with setpoint and adjustment via effectors, which allows it to remain in a particular region of state space. (B) This same capacity is scaled up by cellular networks into anatomical homeostasis: morphogenesis is not simply a feedforward emergent process but rather the ability of living systems to adjust and remodel to specific target morphologies. This requires feedback loops at the transcriptional and biophysical levels, which rely on stored information (e.g., bioelectrical pattern memories) against which to minimize error. (C) This is what underlies complex regeneration such as salamander limbs, which can be cut at any position and result in just the right amount and type of regenerative growth that stops when a correct limb is achieved. Such homeostatic systems are examples of simple goal-directed agents. (D) A focus on the size or scale of goals any given system can pursue allows plotting very diverse intelligences on the same graph, regardless of their origin or composition (Levin, 2019). The scale of their goal-directed activity is estimated (collapsed onto one axis of space and one of time, as in Relativity diagrams). Importantly, this way of visualizing the sophistication of agency is a schematic of goal space—it is not meant to represent the spatial extent of sensing or effector range, but rather the scale of events about which they care and the boundary of states that they can possibly represent or work to change. This defines a kind of cognitive light cone (a boundary to any agent’s area of concern); the largest area represents the “now,” with fading efficacy both backward (accessing past events with decreasing reliability) and forward (limited prediction accuracy for future events). Agents are compound entities, composed of (and comprising) other sub- or super-agents each of which has their own cognitive boundary of various sizes. Images by Jeremy Guay of Peregrine Creative. The ability of a system to exert energy to work toward a state of affairs, overcoming obstacles (to the degree that its sophistication allows) to achieve a particular set of substates is very useful for defining Selves because it grounds the question in well-established control theory and cybernetics (i.e., systems “trying to do things” is no longer magical but is well-established in engineering), and provides a natural way of discovering, defining, and altering the preferences of a system. A common objection is: “surely we can’t say that thermostats have goals and preferences?” The TAME framework holds that whatever true goals and preferences are, there must exist primitive, minimal versions from which they evolved and these are, in an important sense, substrate- and scale-independent; simple homeostatic circuits are an ideal candidate for the “hydrogen atom” of goal-directed activity (Rosenblueth et al., 1943; Turner, 2019). A key tool for thinking about these problems is to ask what a truly minimal example of any cognitive capacity would be like, and to think about transitional forms that can be created just below that. It is logically inevitable that if one follows a complex cognitive capacity backward through phylogeny, one eventually reaches precursor versions of that capacity that naturally suggest the (misguided) question “is that really cognitive, or just physics?” Indeed, a kind of minimal goal-directedness permeates all of physics (Feynman, 1942; Georgiev and Georgiev, 2002; Ogborn et al., 2006; Kaila and Annila, 2008; Ramstead et al., 2019; Kuchling et al., 2020a), supporting a continuous climb of the scale and sophistication of goals. Pursuit of goals is central to composite agency and the “many to one” problem because it requires distinct mechanisms (for measurement of states, storing setpoints, and driving activity to minimize the delta between the former and the latter) to be bound together into a functional unit that is greater than its parts. To co-opt a great quote (Dobzhansky, 1973), nothing in biology makes sense except in light of teleonomy (Pittendrigh, 1958; Nagel, 1979; Mayr, 1992; Schlosser, 1998; Noble, 2010, 2011; Auletta, 2011; Ellis et al., 2012). The degree to which a system can evaluate possible consequences of various actions, in pursuit of those goal states, can vary widely, but is essential to its survival. The expenditure of energy in ways that effectively reach specific states despite uncertainty, limitations of capability, and meddling from outside forces is proposed as a central unifying invariant for all Selves—a basis for the space of possible agents. This view suggests a semi-quantitative multi-axis option space that enables direct comparison of diverse intelligences of all sorts of material implementation and origins (Levin, 2019, 2020). Specifically (Figure 4), a “space-time” diagram can be created where the spatio-temporal scale of any agent’s goals delineates that Self and its cognitive boundaries. Note that the distances on Figure 4D represent not first-order capacities such as sensory perception (how far away can it sense), but second-order capacities of the size of goals (humble metabolic hunger-satiety loops or grandiose planetary-scale engineering ambitions) which a given cognitive system is capable of representing and working toward. At any given time, an Agent is represented by a single shape in this space, corresponding to the size and complexity of their possible goal domain. However, genomes (or engineering design specs) map to an ensemble of such shapes in this space because the borders between Self and world, and the scope of goals an agent’s cognitive apparatus can handle, can all shift during the lifetime of some agents—“in software” (another “great transition” marker). All regions in this space can potentially define some possible agent. Of course, additional subdivisions (dimensions) can easily be added, such as the Unlimited Associative Learning marker (Birch et al., 2020) or aspects of Active Inference (Friston and Ao, 2012; Friston et al., 2015b; Calvo and Friston, 2017; Peters et al., 2017). Some agents, like microbes, have minimal memory (Vladimirov and Sourjik, 2009; Lan and Tu, 2016) and can concern themselves only with a very short time horizon and spatial radius—e.g., follow local gradients. Some agents, e.g., a rat have more memory and some forward planning ability (Hadj-Chikh et al., 1996; Raby and Clayton, 2009; Smith and Litchfield, 2010), but are still precluded from, for example, effectively caring about what will happen 2 months hence, in an adjacent town. Some, like human beings, can devote their lives to causes of enormous scale (future state of the planet, humanity, etc.). Akin to Special Relativity, this formalization makes explicit that class of capacities (in terms of representation of classes of goals) that are forever inaccessible to a given agent (demarcating the edge of the “light cone” of its cognition). In general, larger selves (1) are capable of working toward states of affairs that occur farther into the future (perhaps outlasting the lifetime of the agent itself—an important great transition, in the sense of West et al. (2015), along the cognitive continuum); (2) deploy memories further back in time (their actions become less “mechanism” and more decision-making (Balazsi et al., 2011) because they are linked to a network of functional causes and information with larger diameter); and (3) they expend effort to manage sensing/effector activity in larger spaces [from subcellular networks to the extended mind (Clark and Chalmers, 1998; Turner, 2000; Timsit and Gregoire, 2021)]. Overall, increases of agency are driven by mechanisms that scale up stress (Box 1)—the scope of states that an agent can possibly be stressed about (in the sense of pressure to take corrective action). In this framework, stress (as a system-level response to distance from setpoint states), preferences, motivation, and the ability to functionally care about what happens are tightly linked. Homeostasis, necessary for life, evolves into allostasis (McEwen, 1998; Schulkin and Sterling, 2019) as new architectures allow tight, local homeostatic loops to be scaled up to measure, cause, and remember larger and more complex states of affairs (Di Paulo, 2000; Camley, 2018). BOX 1. Stress as the glue of agency. Tell me what you are stressed about and I will know a lot about your cognitive sophistication. Local glucose concentration? Limb too short? Rival is encroaching on your territory? Your limited lifespan? Global disparities in quality of life on Earth? The scope of states that an agent can possibly be stressed by, in effect, defines their degree of cognitive capacity. Stress is a systemic response to a difference between current state and a desired setpoint; it is an essential component to scaling of Selves because it enables different modules (which sense and act on things at different scales and in distributed locations) to be bound together in one global homeostatic loop (toward a larger purpose). Systemic stress occurs when one sub-agent is not satisfied about its local conditions, and propagates its unhappiness outward as hard-to-ignore signals. In this process, stress pathways serve the same function as hidden layers in a network, enabling the system to be more adaptive by connecting diverse modular inputs and outputs to the same basic stress minimization loop. Such networks scale stress, but stress is also what helps the network scale up its agency—a bidirectional positive feedback loop. The key is that this stress signal is unpleasant to the other sub-agents, closely mimicking their own stress machinery (genetic conservation: my internal stress molecule is the same as your stress molecule, which contributes to the same “wiping of ownership” that is implemented by gap junctional connections). By propagating unhappiness in this way (in effect, turning up the global system “energy” which facilitates tendency for moving in various spaces), this process recruits distant sub-agents to act, to reduce their own perception of stress. For example, if an organ primordium is in the wrong location and needs to move, the surrounding cells are more willing to get out of the way if by doing so they reduce the amount of stress signal they receive. It may be a process akin to run-and-tumble for bacteria, with stress as the indicator of when to move and when to stop moving, in physiological, transcriptional, or morphogenetic space. Another example is compensatory hypertrophy, in which damage in one organ induces other cells to take up its workload, growing or taking on new functions if need be (Tamori and Deng, 2014; Fontes et al., 2020). In this way, stress causes other agents to work toward the same goal, serving as an influence that binds subunits across space into a coherent higher Self and resists the “struggle of the parts” (Heams, 2012). Interestingly, stress spreads not only horizontally in space (across cell fields) but also vertically, in time: effects of stress response is one of the things most easily transferred by transgenerational inheritance (Xue and Acar, 2018). Additional implications of this view are that Selves: are malleable (the borders and scale of any Self can change over time); can be created by design or by evolution; and are multi-scale entities that consist of other, smaller Selves (and conversely, scale up to make larger Selves). Indeed they are a patchwork of agents [akin to Theophile Bordeu’s “many little lives” (Haigh, 1976; Wolfe, 2008)] that overlap with each other, and compete, communicate, and cooperate both horizontally (at their own level of organization) and vertically [with their component subunits and the super-Selves of which they are a part (Sims, 2020)]. Another important invariant for comparing diverse intelligences is that they are all solving problems, in some space (Figure 5). It is proposed that the traditional problem-solving behavior we see in standard animals in 3D space is just a variant of evolutionarily more ancient capacity to solve problems in metabolic, physiological, transcriptional, and morphogenetic spaces (as one possible sequential timeline along which evolution pivoted some of the same strategies to solve problems in new spaces). For example, when planaria are exposed to barium, a non-specific potassium channel blocker, their heads explode. Remarkably, they soon regenerate heads that are completely insensitive to barium (Emmons-Bell et al., 2019). Transcriptomic analysis revealed that relatively few genes out of the entire genome were regulated to enable the cells to resolve this physiological stressor using transcriptional effectors to change how ions and neurotransmitters are handled by the cells. Barium is not something planaria ever encounter ecologically (so there should not be innate evolved responses to barium exposure), and cells don’t turn over fast enough for a selection process (e.g., with bacterial persisters after antibiotic exposure). The task of determining which genes, out of the entire genome, can be transcriptionally regulated to return to an appropriate physiological regime is an example of an unconventional intelligence navigating a large-dimensional space to solve problems in real-time (Voskoboynik et al., 2007; Elgart et al., 2015; Soen et al., 2015; Schreier et al., 2017). Also interesting is that the actions taken in transcriptional space (a set of mRNA states) map onto a path in physiological state (the ability to perform many needed functions despite abrogated K+ channel activity, not just a single state). FIGURE 5. FIGURE 5 Open in a new tab Cognitive agents solve problems in diverse spaces. Intelligence is fundamentally about problem-solving, but this takes place not only in familiar 3D space as “behavior” (control of muscle effectors for movement) (A), but also in other spaces in which cognitive systems try to navigate, in order to reach better regions. This includes the transcriptional space of gene expression (B) here schematized for two genes, anatomical morphospace (C) here schematized for two traits, and physiological space (D) here schematized for two parameters. An example (E) of problem-solving is planaria, which placed in barium (causing their heads to explode due to general blockade of potassium channels) regenerate new heads that are barium-insensitive (Emmons-Bell et al., 2019). They solve this entirely novel (not primed by evolutionary experience with barium) stressor by a very efficient traversal in transcriptional space to rapidly up/down regulate a very small number of genes that allows them to conduct their physiology despite the essential K+ flux blockade. (F) The degree of intelligence of a system can be estimated by how effectively they navigate to optimal regions without being caught in a local maximum, illustrated as a dog which could achieve its goal on the other side of the fence, but this would require going around—temporarily getting further from its goal (a measurable degree of patience or foresight of any system in navigating its space, which can be visualized as a sort of energy barrier in the space, inset). Images by Jeremy Guay of Peregrine Creative. The common feature in all such instances is that the agent must navigate its space(s), preferentially occupying adaptive regions despite perturbations from the outside world (and from internal events) that tend to pull it into novel regions. Agents (and their sub- and super-agents) construct internal models of their spaces (Beer, 2014, 2015; Beer and Williams, 2015; Hoffman et al., 2015; Fields et al., 2017; Hoffman, 2017; Prentner, 2019; Dietrich et al., 2020; Prakash et al., 2020), which may or may not match the view of their action space developed by their conspecifics, parasites, and scientists. Thus, the space one is navigating is in an important sense virtual (belonging to some Agent’s self-model), is developed and often modified “on the fly” (in addition to that hardwired by the structure of the agent), and not only faces outward to infer a useful structure of its option space but also faces inward to map its own body and somatotopic properties (Bongard et al., 2006). The lower-level subsystems simplify the search space for the higher-level agent because their modular competency means that the higher-level system doesn’t need to manage all the microstates [a strong kind of hierarchical modularity (Zhao et al., 2006; Lowell and Pollack, 2016)]. In turn, the higher-level system deforms the option space for the lower-level systems so that they do not need to be as clever, and can simply follow local energy gradients. The degree of intelligence, or sophistication, of an agent in any space is roughly proportional to its ability to deploy memory and prediction (information processing) in order to avoid local maxima. Intelligence involves being able to temporarily move away from a simple vector toward one’s goals in a way that results in bigger improvements down the line; the agent’s internal complexity has to facilitate some degree of complexity (akin to hidden layers in an artificial neural network which introduce plasticity between stimulus and response) in the goal-directed activity that enables the buffering needed for patience and indirect paths to the goal. This buffering enables the flip side of homeostatic problem-driven (stress reduction) behavior by cells: the exploration of the space for novel opportunities (creativity) by the collective agent, and the ability to acquire more complex goals [in effect, beginning the climb to Maslow’s hierarchy (Taormina and Gao, 2013)]. Of course it must be pointed out that this way of conceiving intelligence is one of many, and is proposed here as a way to enable the concept to be experimentally ported over to unfamiliar substrates, while capturing what is essential about it in a way that does not depend on arbitrary restrictions that will surely not survive advances in synthetic bioengineering, machine learning, and exobiology. Another important aspect of intelligence that is space-agnostic is the capacity for generalization. For example, in the barium planaria example discussed above, it is possible that part of the problem-solving capacity is due to the cells’ ability to generalize in physiological space. Perhaps the cells recognize the physiological stresses induced by the novel barium stimulus as a member of the wider class of excitotoxicity induced by evolutionarily-familiar epileptic triggers, enabling them to deploy similar solutions (in terms of actions in transcriptional space). Such abilities to generalize have now been linked to measurement invariance (Frank, 2018), showing its ancient roots in the continuum of cognition. Consistent with the above discussion, complex agents often consist of components that are themselves competent problem-solvers in their own (usually smaller, local) spaces. The relationship between wholes and their parts can be as follows. An agent is an integrated holobiont to the extent that it distorts the option space, and the geodesics through it, for its subunits (perhaps akin to how matter and space affect each other in general relativity) to get closer to a high-level goal in its space. A similar scheme is seen in neuroscience, where top-down feedback helps lower layer neurons to choose a response to local features by informing them about more global features (Krotov, 2021). At the level of the subunits, which know nothing of the higher problem space, this simply looks like they are minimizing free energy and passively doing the only thing they can do as physical systems: this is why if one zooms in far enough on any act of decision-making, all one ever sees is dumb mechanism and “just physics.” The agential perspective (Godfrey-Smith, 2009) looks different at different scales of observation (and its degree is in the eye of a beholder who seeks to control and predict the system, which includes the Agent itself, and its various partitions). This view is closely aligned with that of “upper directedness” (McShea, 2012), in which the larger system directs its components’ behavior by constraints and rewards for coarse-grained outcomes, not microstates (McShea, 2012). Note that these different competing and cooperating partitions are not just diverse components of the body (cells, microbiome, etc.) but also future and past versions of the Self. For example, one way to achieve the goal of a healthier metabolism is to lock the refrigerator at night and put the keys somewhere that your midnight self, which has a shorter cognitive boundary (is willing to trade long-term health for satiety right now) and less patience, is too lazy to find. Changing the option space, energy barriers, and reward gradients for your future self is a useful strategy for reaching complex goals despite the shorter horizons of the other intelligences that constitute your affordances in action space. The most effective collective intelligences operate by simultaneously distorting the space to make it easy for their subunits to do the right thing with no comprehension of the larger-scale goals, but themselves benefit from the competency of the subunits which can often get their local job done even if the space is not perfectly shaped (because they themselves are homeostatic agents in their own space). Thus, instances of communication and control between agents (at the same or different levels) are mappings between different spaces. This suggests that both evolution’s, and engineers’, hard work is to optimize the appropriate functional mapping toward robustness and adaptive function. Next, we consider a practical example of the application of this framework to an unconventional example of cognition and flexible problem-solving: morphogenesis, which naturally leads to specific hypotheses of the origin of larger biological Selves (scaling) and its testable empirical (biomedical) predictions (Dukas, 1998). This is followed with an exploration of the implications of these concepts for evolution, and a few remarks on consciousness. Somatic Cognition: An Example of Unconventional Agency in Detail “Again and again terms have been used which point not to physical but to psychical analogies. It was meant to be more than a poetical metaphor…” – Spemann (1967) An example of TAME applied to basal cognition in an unconventional substrate is that of morphogenesis, in which the mechanisms of cognitive binding between subunits are now partially known, and testable hypotheses about cognitive scaling can be formulated [explored in detail in Friston et al. (2015a) and Pezzulo and Levin (2015, 2016)]. It is uncontroversial that morphogenesis is the result of collective activity: individual cells work together to build very complex structures. Most modern biologists treat it as clockwork [with a few notable exceptions around the recent data on cell learning (di Primio et al., 2000; Brugger et al., 2002; Norman et al., 2013; Yang et al., 2014; Stockwell et al., 2015; Urrios et al., 2016; Tweedy and Insall, 2020; Tweedy et al., 2020)], preferring a purely feed-forward approach founded on the idea of complexity science and emergence. On this view, there is a privileged level of causation—that of biochemistry—and all of the outcomes are to be seen as the emergent consequences of highly parallel execution of local rules (a cellular automaton in every sense of the term). Of course, it should be noted that the forefathers of developmental biology, such as Spemann (1967), were already well-aware of the possible role of cognitive concepts in this arena and others have occasionally pointed out detailed homologies (Grossberg, 1978; Pezzulo and Levin, 2015). This becomes clearer when we step away from the typical examples seen in developmental biology textbooks and look at some phenomena that, despite the recent progress in molecular genetics, remain important knowledge gaps (Figure 6). FIGURE 6. FIGURE 6 Open in a new tab Morphogenesis as an example of collective intelligence and plasticity. The results of complex morphogenesis are the behavior in morphospace of a collective intelligence of cells. It is essential to understand this collective intelligence because by themselves, progress in molecular genetics is insufficient. For example, despite genomic information and much pathway data on the behavior of stem cells in planarian regeneration, there are no models predicting what happens when cells from a flat-headed species are injected into a round-headed species (A): what kind of head will they make, and will regeneration/remodeling ever stop, since the target morphology can never match what either set of cells expects? Development has the ability to overcome unpredictable perturbations to reach its goals in morphospace: tadpoles made with scrambled positions of craniofacial organs can make normal frogs (B) because the tissues will move from their abnormal starting positions in novel ways until a correct frog face is achieved (Vandenberg et al., 2012). This illustrates that the genetics seeds the development of hardware executing not an invariant set of movements but rather an error minimization (homeostatic) loop with reference to a stored anatomical setpoint (target morphology). The paths through morphospace are not unique, illustrated by the fact that when frog legs are induced to regenerate (C), the intermediate stages are not like the developmental path of limb development (forming a paddle and using programmed cell death to separate the digits) but rather like a plant (C′), in which a central core gives rise to digits growing as offshoots (green arrowheads) which nevertheless ends up being a very normal-looking frog leg (Tseng and Levin, 2013). (D) The plasticity extends across levels: when newt cells are made very large by induced polyploidy, they not only adjust the number of cells that work together to build kidney tubules with correct lumen diameter, but can call up a completely different molecular mechanism (cytoskeletal bending instead of cell:cell communication) to make a tubule consisting in cross-section of just 1 cell wrapped around itself; this illustrates intelligence of the collective, as it creatively deploys diverse lower-level modules to solve novel problems. The plasticity is not only structural but functional: when tadpoles are created to (E) have eyes on their tails (instead of in their heads), the animals can see very well (Blackiston and Levin, 2013), as revealed by their performance in visual learning paradigms (F). Such eyes are also competent modules: they first form correctly despite their aberrant neighbors (muscle, instead of brain), then put out optic nerves which they connect to the nearby spinal cord, and later they ignore the programmed cell death of the tail, riding it backward to end up on the posterior of the frog (G). All of this reveals the remarkable multi-scale competency of the system which can adapt to novel configurations on the fly, not requiring evolutionary timescales for adaptive functionality (and providing important buffering for mutations that make changes whose disruptive consequences are hidden from selection by the ability of modules to get their job done despite changes in their environment). Panels (A,C’,D) are courtesy of Peregrine Creative. Panel (C) is from Xenbase and Aisun Tseng. Panels (E–G) are courtesy of Douglas Blackiston. Panel (B) is used with permission from Vandenberg et al. (2012), and courtesy of Erin Switzer. Goal-Directed Activity in Morphogenesis Morphogenesis (broadly defined) is not only a process that produces the same robust outcome from the same starting condition (development from a fertilized egg). In animals such as salamanders, cells will also re-build complex structures such as limbs, no matter where along the limb axis they are amputated, and stop when it is complete. While this regenerative capacity is not limitless, the basic observation is that the cells cooperate toward a specific, invariant endstate (the target morphology), from diverse starting conditions, and cease their activity when the correct pattern has been achieved. Thus, the cells do not merely perform a rote set of steps toward an emergent outcome, but modify their activity in a context-dependent manner to achieve a specific anatomical target morphology. In this, morphogenetic systems meet James’ test for minimal mentality: “fixed ends with varying means” (James, 1890). For example, tadpoles turn into frogs by rearranging their craniofacial structures: the eyes, nostrils, and jaws move as needed to turn a tadpole face into a frog face (Figure 6B). Guided by the hypothesis that this was not a hardwired but an intelligent process that could reach its goal despite novel challenges, we made tadpoles in which these organs were in the wrong positions—so-called Picasso Tadpoles (Vandenberg et al., 2012). Amazingly, they tend to turn into largely normal frogs because the craniofacial organs move in novel, abnormal paths [sometimes overshooting and needing to return a bit (Pinet et al., 2019)] and stop when they get to the correct frog face positions. Similarly, frog legs that are artificially induced to regenerate create a correct final form but not via the normal developmental steps (Tseng and Levin, 2013). Students who encounter such phenomena and have not yet been inoculated with the belief that molecular biology is a privileged level of explanation (Noble, 2012) ask the obvious (and proper) question: how does it know what a correct face or leg shape is? Examples of remodeling, regulative development (e.g., embryos that can be cut in half and produce normal monozygotic twins), and regeneration, ideally illustrate the goal-directed nature of cellular collectives. They pursue specific anatomical states that are much larger than any individual cells and solve problems in morphospace in a context-sensitive manner—any swarm of miniature robots that could do this would be called a triumph of collective intelligence in the engineering field. Guided by the TAME framework, two questions come within reach. First, how does the collective measure current state and store the information about the correct target morphology? Second, if morphogenesis is not at the clockwork level on the continuum of persuadability but perhaps at that of the thermostat, could it be possible to re-write the setpoint without rewiring the machine (i.e., in the context of a wild-type genome)? Pattern Memory: A Key Component of Homeostatic Loops Deer farmers have long known of trophic memory: wounds made on a branched antler structure in 1 year, will result in ectopic tines growing at that same location in subsequent years, long after the original rack of antlers has fallen off (Bubenik and Pavlansky, 1965; Bubenik, 1966; Lobo et al., 2014). This process requires cells at the growth plate in the scalp to sense, and remember for months, the location of a transient damage event within a stereotypical branched structure, and reproduce it in subsequent years by over-riding the wild-type stereotypical growth patterns of cells, instead guiding them to a novel outcome. This is an example of experience-dependent, re-writable pattern memory, in which the target morphology (the setpoint for anatomical homeostasis) is re-written within standard hardware. Planarian flatworms can be cut into multiple pieces, and each fragment regenerates precisely what is missing at each location (and re-scales the remaining tissue as needed) to make a perfect little worm (Cebrià et al., 2018). Some species of planaria have an incredibly messy genome—they are mixoploid due to their method of reproduction: fission and regeneration, which propagates any mutations that don’t kill the stem cell and expands it throughout the lineage [reviewed in Fields et al. (2020)]. Despite the divergence of genomic information, the worms are champion regenerators, with near 100% fidelity of anatomical structure. Recent data have identified one set of mechanisms mediating the ability of the cells to make, for example, the correct number of heads: a standing bioelectrical distribution across the tissue, generated by ion channels and propagated by electrical synapses known as gap junctions (Figures 7A–D). Manipulation of the normal voltage pattern by targeting the gap junctions (Sordillo and Bargmann, 2021) or ion channels can give rise to planaria with one, two, or 0 heads, or heads with shape (and brain shape) resembling other extant species of planaria (Emmons-Bell et al., 2015; Sullivan et al., 2016). Remarkably, the worms with abnormal head number are permanently altered to this pattern, despite their wild-type genetics: cut into pieces with no further manipulations, the pieces continue to regenerate with abnormal head number (Oviedo et al., 2010; Durant et al., 2017). Thus, much like the optogenetic techniques used to incept false behavioral memories into brains (Vetere et al., 2019), modulation of transient bioelectric state is a conserved mechanism by which false pattern memories can be re-written into the genetically-specified electrical circuits of a living animal. FIGURE 7. FIGURE 7 Open in a new tab Bioelectrical pattern memories. Planarian fragments reliably regenerate whatever is missing, and stop when a correct worm is complete. Normal planaria (A) have 1 head and 1 tail (A-1), expression of anterior genes in the head (A-2), and a standing pattern of resting potential that depolarized on the end that should make a head [(A-3), revealed by voltage-reporting fluorescent dye, depolarized region marked with orange arrowhead]. When a middle portion is amputated, it regenerates to a correct 1-headed worm (A-4). It is possible to edit the information structure which encodes the target morphology (shape to which the fragments will regenerate). In worms that are anatomically normal (A′-1), with normal gene expression (A′-2), their bioelectric pattern can be altered in place [(A′-3), orange arrowheads mark the two depolarized ends] using ion channel-targeting drugs or RNAi. The result, after injury, will be a fully viable 2-headed worm (A′-4). Importantly, the pattern shown in panel (A′-3) is not a voltage map of the final 2-headed worm: it’s a map of a 1-headed animal before cutting, which already has the induced false memory indicating that a correct worm should have 2 heads. In other words, the bioelectrics can diverge from the current state—it is not simply a readout of what the anatomy is doing now, but an orthogonal information medium that is used to guide future changes of anatomy. This information is latent, only guiding the cellular collective’s anatomical homeostasis activity after injury. Thus it is also a basal example of counterfactual representation, referring to what should happen if an injury occurs, not what is happening now. Such changes to the bioelectric target morphology are true memories because they are re-writable but also long-term stable: if cut again, in water with no more channel-perturbing reagents, multiple rounds of regeneration of a genetically wild-type worm continue to give rise to 2-headed forms (B), which can be re-set back to normal by a different bioelectric perturbation (Oviedo et al., 2010). The control of morphology by bioelectric patterns is mediated as in the brain (C) by cells which have ion channels that set resting potential across the membrane (Vmem) and propagate those states in computational networks to their neighbors, via electrical synapses known as gap junctions. All cells, not just neurons (D) do this, and bioelectric signaling is an ancient information processing modality that pre-dates neurons and brains (Fields et al., 2020; Levin, 2021a). The ability of voltage states to functionally specify modular anatomy is seen when an ion channel is used to set membrane voltage in endodermal cells fated to be gut, to an eye-like bioelectric prepattern (E), which then create an eye on the gut (red arrowhead) (Pai et al., 2012). This phenomenon has 2 levels of instruction (F): in addition to our use of voltage to instruct shape at the organ level (not micromanaging individual eye components), the ion channel mRNA-injected cells (cyan β-galactose marker) further instruct their neighbors (brown cells) to participate in forming this ectopic lens. Images in panels (C,D) are courtesy of Peregrine Creative. Images in panels (A,A′) are taken with permission from Durant et al. (2017). Embryo image in panel (E) is from Xenbase. Panel (F) is used with permission from Zahn et al. (2017). Multi-Scale Competency of Growth and Form A key feature of morphogenesis is that diverse underlying molecular mechanisms can be deployed to reach the same large-scale goal. This plasticity and coarse-graining over subunits’ states is a hallmark of collective cognition, and is also well known in neuroscience (Prinz et al., 2004; Otopalik et al., 2017). Newt kidney tubules normally have a lumen of a specific size and are made up (in cross section) of 8–10 cells (Fankhauser, 1945a,b). When the cell size is experimentally enlarged, the same tubules are made of a smaller number of the bigger cells. Even more remarkable than the scaling of the cell number to unexpected size changes (on an ontogenetic, not evolutionary, timescale) is the fact that if the cells are made really huge, just one cell wraps around itself and still makes a proper lumen (Figure 6D). Instead of the typical cell-cell interactions that coordinate tubule formation, cytoskeletal deformations within one cell can be deployed to achieve the same end result. As in the brain, the levels of organization exhibit significant autonomy in the details of their molecular activity but are harnessed toward an invariant system-level outcome. Specific Parallels Between Morphogenesis and Basal Cognition The plasticity of morphogenesis is significantly isomorphic to that of brains and behavior because the communication dynamics that scale individual neural cells into a coherent Self are ones that evolution honed long before brains appeared, in the context of morphogenic control (Fields et al., 2020), and before that, in metabolic control in bacterial biofilms (Prindle et al., 2015; Liu et al., 2017; Martinez-Corral et al., 2019; Yang et al., 2020). Each genome specifies cellular hardware that implements signaling circuits with a robust, reliable default “inborn” morphology—just as genomes give rise to brain circuits that drive instinctual behavior in species that can build nests and do other complex things with no training. However, evolution selected for hardware that can be reprogrammed by experiences, in addition to its robust default functional modes—in body structure, as well as in brain-driven behavior. Many of the brain’s special features are to be found, unsurprisingly, in other forms outside the central nervous system. For example, mirror neurons and somatotopic representation are seen in limbs’ response to injury, where the type and site of damage to one limb can be read out within 30 s from imaging the opposite, un-injured limbs (Busse et al., 2018). Table 2 shows the many parallels between morphogenetic and cognitive systems. TABLE 2. isomorphism between cognition and pattern formation. Cognitive concept Morphogenetic concept Patterns of activation across neural networks processing information Differential patterns of Vmem across tissue formed by propagation of bioelectric states through gap junction synapses. Local field potential (EEG) Vmem distribution of cell group Intrinsic plasticity Change of ion channel expression based on Vmem levels Synaptic plasticity Change of cell:cell connectivity via Vmem’s regulation of gap junctional connectivity Activity-dependent transcriptional changes Bioelectric signals’ regulating gene expression during patterning Neuromodulation, and neurotransmitters controlled by electrical dynamics to regulate genes in neurons Developmental (pre-nervous) signaling via the same neurotransmitters (e.g., serotonin) moving under control of bioelectrical gradients to regulate second messenger pathways and gene expression. Direct transmission Cell:cell sharing of voltage via nanotubes or gap junctions Volume transmission Cell:cell communication via ion levels outside the membrane or voltage-dependent neurotransmitter release Synaptic Vesicles Exosomes Sensitization Cells become sensitized stimuli, such as for example to BMP antagonists during development Functional lateralization Left-right asymmetry of body organs Taste and olfactory perception Morphogenetic signaling by diffusible biochemical ligands Activity-dependent modification of CNS Control of anatomy by bioelectric signaling within those same cells Critical plasticity periods Competency windows for developmental induction events Inborn behaviors (instincts) Emergent morphogenetic cascades as “default” outcomes of a genetically-specified bioelectric hardware—hardwired patterning programs (mosaic development Voluntary movement Remodeling, regeneration, metamorphosis Memory Short range: epigenetic cell memory Medium range: Regeneration of specific body organs. Long range: Morphological homeostasis over decades as individual cells senesce; altering basic body anatomy in planaria by direct manipulation of bioelectric circuit Counterfactual memories Ability of 1-headed planarian bodies to store bioelectric patterns indicative of 1-headed or 2-headed forms, which are latent memories that become instructive upon damage to the organism. Perceptual Bistability Cryptic Planaria, induced by gap-junctional disruption, fragments of which stochastically regenerate as 1-headed or 2-headed forms, shifting between two different bioelectrical representations of a target morphology (pattern memory). Edge detection in retina Sharp boundaries between regions of different Vmem induce downstream gene expression and morphogenetic outcomes Pattern completion ability of neural networks (e.g., attractor nets) Regeneration of missing parts in partial fragments (e.g., planaria, salamander appendages, etc.) Forgetting Degradation of target morphology setpoint information leading to cancer and loss of regenerative ability Addiction Dependency on cellular signals, such as nerve addiction in limb regeneration and cancer addiction to specific molecules. Encoding Representation of patterning goal states by bioelectric properties of tissue Visual system feature detection Organ-level monitoring of body configuration and detection of specific boundaries by tissue (such as the Vmem boundary that drives brain morphogenesis) Holographic (distributed) storage Any small piece of a planarian remembers the correct pattern (even if it has been re-written) Behavioral plasticity Regulative developmental programs and regenerative capacity Self-modeling Representations of current and future morphogenetic states by bioelectric patterns such as the planarian prepattern or the bioelectric face pattern in vertebrates Goal-seeking Embryogenesis and regeneration work toward a specific target configuration despite perturbations Adaptivity and Intelligence Morphological rearrangements carry out novel, not hardwired, movements to reach the same anatomical configuration despite unpredictable initial starting state Age-dependent cognitive decline Age-dependent loss of regenerative ability Top-down control Place conditioning for drug effects—top-down control of signaling pathways Open in a new tab Possible mapping of concepts in cognitive neuroscience to examples in pattern formation. Not Just Philosophy: Why These Parallels Matter The view of anatomical homeostasis as a collective intelligence is not a neutral philosophical viewpoint—it makes strong predictions, some of which have already borne fruit. It led to the discovery of reprogrammable head number in planaria (Nogi and Levin, 2005) and of pre-neural roles for serotonin (Fukumoto et al., 2005a,b). It explains the teratogenicity for pre-neural exposure to ion channel or neurotransmitter drugs (Hernandez-Diaz and Levin, 2014), the patterning defects observed in human channelopathies in addition to the neurological phenotypes [reviewed in Srivastava et al. (2020)], and the utility of gap junction blockers as general anesthetics. Prediction derived from the conservation and scaling hypotheses of TAME can be tested via bioinformatics. Significant and specific overlap are predicted for genes involved in morphogenesis and cognition (categories of memory and learning). This is already known for ion channels, connexin (gap junction) genes, and neurotransmitter machinery, but TAME predicts a widespread re-use of the same molecular machinery. Cell-cell communication and cellular stress pathways should be involved in internal conflict between psychological modules (Reinders et al., 2019) and social behavior, while memory genes should be identified in genetic investigations of cancer, regeneration, and embryogenesis. Another key prediction that remains to be tested (ongoing in our lab) is trainability of morphogenesis. The collective intelligence of tissues could be sophisticated enough to be trainable via reinforcement learning for specific morphological outcomes. Learning has been suggested by clinical data in the heart (Zoghi, 2004), bone (Turner et al., 2002; Spencer and Genever, 2003), and pancreas (Goel and Mehta, 2013). It is predicted that using rewards and punishments (with nutrients/endorphins and shock), not micromanagement of pathway hardware, could be a path to anatomical control in clinical settings, whether for morphology or for gene expression (Biswas et al., 2021). This would have massive implications for regenerative medicine, because the complexity barrier prevents advances such as genomic editing from impacting e.g., limb regeneration in the foreseeable future. The same reasons for which we would train a rat for a specific behavior, rather than control all of the relevant neurons to force it to do it like a puppet, explain why the direct control of molecular hardware is a far more difficult biomedical path than understanding the sets of stimuli that could motivate tissues to build a specific desired structure. The key lesson of computer science has been that even with hardware we understand (if we built it ourselves), it is much more efficient and powerful to understand the software and evince desired outcomes by the appropriate stimulation and signaling, not physical rewiring. If the hardware is reprogrammable (and it is here argued that much of the biological hardware meets this transition), one can offload much of the complexity onto the system itself, taking advantage of whatever competence the sub-modules have. Indeed, neuroscience itself may benefit from cracking a simpler version of the problem, in the sense of neural decoding, done first in non-neural tissues. Non-neural Bioelectricity: What Bodies Think About The hardware of the brain consists of ion channels which set the cells’ electrical state, and controllable synapses (e.g., gap junctions) which can propagate those states across the network. This machinery, including the neurotransmitters that eventually transduce these computations into transcriptional and other cell behaviors, is in fact highly conserved and present in all cells, from the time of fertilization (Figures 7C,D). A major difference between neural and non-neural bioelectricity is the time constant with which it acts [brains speed up the system into millisecond scales, while developmental voltage changes occur in minutes or hours (Harris, 2021; Levin, 2021a)]. Key aspects of this system in any tissue that enable it to support flexible software include the fact that both ion channels and gap junctions are themselves voltage sensitive—in effect, they are transistors (voltage-gated current conductances). This enables evolution to exploit the laws of physics to rapidly generate very complex circuits with positive (memory) and negative (robustness) feedback (Law and Levin, 2015; Cervera et al., 2018, 2019a,b,2020a). The fact that a transient voltage state passing through a cell can set off a cycle of progressive depolarization (like an action potential) or gap junctional (GJ) closure means that such circuits readily form dynamical systems memories which can store different information and change their computational behavior without changing the hardware (i.e., not requiring new channels or gap junctions) (Pietak and Levin, 2017); this is obvious in the action potential propagations in neural networks but is rarely thought about in development. It should be noted that there are many additional biophysical modalities, such as parahormones, volume conduction, biomechanics (strain and other forces), cytoskeletal dynamics, and perhaps even quantum coherence events that could likewise play interesting roles. These are not discussed here only due to length limitations; instead, we are focusing on the bioelectric mechanisms as one particularly illustrative example of how evolution exploits physics for computation and cognition. Consistent with its proposed role, slowly-changing resting potentials serve as instructive patterns guiding embryogenesis, regeneration, and cancer suppression (Bates, 2015; Levin et al., 2017; McLaughlin and Levin, 2018). In addition to the pattern memories encoded electrically in planaria (discussed above), bioelectric prepatterns have also been shown to dictate the morphogenesis of the face, limbs, and brain, and function in determining primary body axes, size, and organ identity [reviewed in Levin and Martyniuk (2018)]. One of the most interesting aspects of developmental bioelectricity is its modular nature: very simple voltage states trigger complex, downstream patterning cascades. As in the brain, modularity goes hand-in-hand with pattern completion: the ability of such networks to provide entire behaviors from partial inputs. For example, Figure 7F shows how a few cells transduced with an ion channel that sets them into a “make the eye here” trigger recruit their neighbors, in any region of the body, to fulfill the purpose of the subroutine call and create an eye. Such modularity makes it very easy for evolution to develop novel patterns by re-using powerful triggers. Moreover, as do brains, tissues use bioelectric circuits to implement pattern memories that set the target morphology for anatomical homeostasis (as seen in the planarian examples above). This reveals the non-neural material substrate that stores the information in cellular collectives, which is a distributed, dynamic, re-writable form of storage that parallels recent discoveries of how group knowledge is stored in larger-scale agents such as animal swarms (Thierry et al., 1995; Couzin et al., 2002). Finally, bioelectric domains (Pai et al., 2017, 2018; Pitcairn et al., 2017; McNamara et al., 2019, 2020) set the borders for groups of cells that are going to complete a specific morphogenetic outcome—a system-level process like “make an eye.” They define the spatio-temporal borders of the modular activity, and suggest a powerful model for how Selves scale in general. A Bioelectric Model of the Scaling of the Self Gap junctional connections between cells provide an interesting case study for how the borders of the Self can expand or contract, in the case of a morphogenetic collective intelligence (Figure 8). Crucially, gap junctions [and gap junctions extended by tunneling nanotubes (Wang et al., 2010; Ariazi et al., 2017)] enable a kind of cellular parabiosis—a regulated fusion between cells that enables lateral inheritance of physiological information, which speeds up processing in the same way that lateral gene inheritance potentiates change on evolutionary timescales. The following is a case study hypothesizing one way in which evolution solves the many-into-one problem (how competent smaller Selves bind into an emergent higher Self), and how this process can break down leading to a reversal (shrinking) of the Self boundary (summarized in Table 3). FIGURE 8. FIGURE 8 Open in a new tab Scaling of computation in cells. Individual cells (A) have a degree of computational capacity consisting of the ability to sense local microenvironment, and some memory and ability to anticipate into the future. When assembling into networks (A′), tissues acquire the ability to sense and act at greater spatial distance, as well as gain larger capacity for memory and prediction via greater computational capacity. As neural networks use hidden layers to abstract patterns in data and recognize meso-scale features (B), tissue networks gain the capacity to represent information larger than the molecular and cell level: each cell’s activity (differentiation, migration, etc.) can be the result of other layers of cells processing information about current and past states, enabling decision-making with respect to tissue, organ, or whole organism-scale anatomy. (C) Much as some neural networks store individual memories as attractors in their state space, bioelectric circuits’ attractors function as pattern memories, triggering cells to execute behaviors that implement anatomical outcomes like number and location of heads in planaria. Images courtesy of Jeremy Guay of Peregrine Creative. TABLE 3. An example of the scaling of cognition. •Each Self has a cognitive capacity defined by the spatial, temporal, and complexity metrics on the goals it can possibly pursue. •Biological Selves scale up by cells’ joining into computational networks that can pursue larger-scale (anatomical, not just metabolic) goals. •Networks increase the spatial reach of sensing and actuation, and increase the computational capacity which allows scaling up of goals and of the states that can induce stress. •Bodies consist of components which are themselves competent (goal-seeking modules that navigate their own spaces) and can achieve specific outcomes despite perturbations and changing conditions. •Gap junctions are a unique scaling mechanism which, by linking cells’ internal milieus, wipes ownership information on signaling molecules. This partially erases the informational identity of the cellular subunits, driving up cooperation and resulting in novel tissue and organ-level Selves with morphological-scale goals. •Bioelectric networks underlie the computations of cell collectives at the tissue, organ, and organism scale, propagating stress information, state sensing, and morphogenetic instructive cues over larger areas. •Selves can dissociate (scale down), as occurs in cancer, by shrinking the computational boundaries of some subunits that de-couple from the network. Open in a new tab Single cells (e.g., the protozoan Lacrymaria olor) are very competent in handling morphological, physiological, and behavioral goals on the scale of one cell. When connected to each other via gap junctions, as in metazoan embryos, several things happen (much of which is familiar to neuroscientists and workers in machine learning in terms of the benefits of neural networks) that lead to the creation of a Self with a new, larger cognitive boundary. First, when cells join into an electrochemical network, they can now sense events, and act, on a much larger physical “radius of concern” than a single cell. Moreover, the network can now integrate information coming from spatially disparate regions in complex ways that result in activity in other spatial regions. Second, the network has much more computational power than any of its individual cells (nodes), providing an IQ boost for the newly formed Self. In such networks, Hebbian dynamics on the electrical synapse (GJ) can provide association between action in one location and reward in another, which enables the system to support credit assignment at the level of the larger individual. The third consequence of GJ connectivity is the partial dissolution of informational boundaries between the subunits. GJ-mediated signals are unique because they give each cell immediate access to the internal milieu of other cells. A conventional secreted biochemical signal arrives from the outside, and when it triggers cell receptors on the surface, the cell clearly knows that this information originated externally (and can be attended to, ignored, etc.)—it is easy to maintain boundary between Self and world. However, imagine a signal like a calcium spike originating in a cell due to some damage stimulus for example. When that calcium propagates onto the GJ-coupled neighbor, there are no metadata on that signal marking its origin; the recipient cell only knows that a calcium transient occurred, and cannot tell that this information does not belong to it. The downstream effects of this second messenger event are a kind of false memory for the recipient cell, but a true memory for the collective network of the stimulus that occurred in one part of the individual. This wiping of ownership information for GJ signals as they propagate through the network is critical to enabling a partial “mind meld” between the cells: keeping identity (in terms of distinct individual history of physiological states—memory) becomes very difficult, as small informational molecules propagate and mix within the network. Thus, this property of GJ coupling promotes the creation of a larger Self by partially erasing the mnemic boundaries between the parts which might impair their ability to work toward a common goal. This is a key part of the scaling of the Self by enlarging toward common goals—not by micromanagement, but by bringing multiple subunits into the same goal-directed loop by tightly coupling the sensing, memory, and action steps in a syncytium where all activity is bound toward a system-level teleonomic process. When individual identities are blurred in favor of longer time-scale, larger computations in tissues, small-horizon (myopic) action in individual cells (e.g., cancer cells’ temporary gains followed by maladaptive death of the host) leads to a more adaptive longer-term future as a healthy organism. In effect, this builds up long-term collective rationality from the action of short-sighted irrational agents (Sasaki and Biro, 2017; Berdahl et al., 2018). It is important to note that part of this story has already been empirically tested in assays that reveal the shrinking as well as the expansion of the Self boundary (Figure 9). One implication of these hypotheses is that the binding process can break down. Indeed this occurs in cancer, where oncogene expression and carcinogen exposure leads to a closure of GJs (Vine and Bertram, 2002; Leithe et al., 2006). The consequence of this is transformation to cancer, where cells revert to their ancient unicellular selves (Levin, 2021b)—shrinking their computational boundaries and treating the rest of the body as external environment. The cells migrate at will and proliferate as much as they can, fulfilling their cell-level goals—metastasis [but also sometimes attempting to, poorly, reboot their multicellularity and make tumors (Egeblad et al., 2010)]. The model implies that this phenotype can be reverted by artificially managing the bioelectric connections between a cell and its neighbors. Indeed, recent data show that managing this connectivity can override default genetically-determined states, inducing metastatic melanoma in a perfectly wild-type background (Blackiston et al., 2011) or suppressing tumorigenesis induced by strong oncogenes like p53 or KRAS mutations (Chernet and Levin, 2013a,b). The focus on physiological connectivity (information dynamics)—the software—is consistent with the observed facts that genetic alterations (hardware) are not necessary to either induce or revert cancer [reviewed in Chernet and Levin (2013a)]. FIGURE 9. FIGURE 9 Open in a new tab Gap junctions and the cellular collective. Communication via diffusible and biomechanical signals can be sensed by receptors at the membrane as messages coming from the outside of a cell (A). In contrast, cells coupled by gap junctions enable signals to pass directly from one cell’s internal milieu into another. This forms a partial syncytium which helps erase informational boundaries between cells, as memory molecules (results of pathway dynamics) propagate across such cell groups without metadata on which cell originated them. The versatile gating of GJ synapses allows the formation of multicellular Selves that own memories of physiological past events at the tissue level (not just individual cells’) and support larger target patterns, enabling them to cooperate to make complex organs (B). This process can break down: when oncogenes are expressed in tadpoles, voltage dye imaging (C) reveals the abnormal voltage state of cells that are disconnected bioelectrically from their neighbors, reverting to an ancient unicellular state (metastasis) that treats the rest of the body as external environment and grows out of control as tumors (D). This process can be prevented (Chernet and Levin, 2013a,b; Chernet et al., 2016) by artificially regulating their bioelectric state [e.g., co-injecting a hyperpolarizing channel with the oncogene, (E)]. In this case the tissue forms normally [(F), green arrow], despite the very strong presence of the oncogene [(G), red label]. This illustrates the instructive capacity of bioelectric networks to dominate single cell and genetic states to control large-scale tissue outcomes. Panels (A,A′,B) courtesy of Jeremy Guay of Peregrine Creative. Panels (C–D) are used with permission from Chernet and Levin (2013a). Panels (E–G) used with permission from Chernet and Levin (2013b). All these dynamics lead to a few interesting consequences. GJ-mediated communications are not merely conversations (in the way that external signaling is)—they are binding, in the sense that once a GJ is open, a cell is subject to whatever comes in from the neighbor. In the same sense, having a synapse makes one vulnerable to the state of neighbors. GJs spread (dilute) the pain of depolarization, but at the same time give a cell’s neighbors the power to change its state. Compatible with the proposal that the magnitude of a Self is the scale and complexity of states by which it can be stressed, connections by tunable, dynamic GJs greatly expand the spatial, temporal, and complexity of things that can irritate cells; complex events from far away can now percolate into a cell via non-linear GJ paths through the network, and enabling the drive to minimize such events now necessarily involves homeostatic activity of goal states, sensing, and activity on a much larger scale. Stress signals, propagating through such networks, incentivize other regions of the tissue to act cooperatively in response to distant events by harnessing their selfish drive to reduce their own stress. This facilitates the coherent, system-level response to perturbations beyond their local consequences, and gives rise to larger Selves able to react coherently to stressful departures from more complex, spatially-distributed allostatic setpoints. For example, whereas a solitary cell might be stressed (and react to) abnormal local pH, cells that are part of a transplanted salamander limb will be induced to a more grandiose activity: they will change the number of fingers they produce to be correct relative to the limb’s new position in the host’s body (Ruud, 1929), a decision that involves large-scale sensing, decision-making, and control. A fourth consequence of the coupling is that cooperation in general is greatly enhanced. In the game theory sense, it is impossible to cheat against your neighbor if you are physiologically coupled. Any positive or negative effects of a cell’s actions toward the neighboring cell are immediately propagated back to it, in effect producing one individual in which the parts cannot “defect” against each other. This dynamic suggests an interesting twist on Prisoners’ Dilemma models in which the number of agents is not fixed, because they have the options of Cooperate, Defect, Merge, and Split (we are currently analyzing such models). Specifically, merging with another agent creates an important dimensionality reduction (because defection is no longer an option); this not only changes the calculus of game theory as applied to biological interactions, but also the action space itself. These dynamics take place on a developmental timescale, complementing the rich existing literature on game theory in evolution (Maynard Smith and Szathmáry, 1995; Maynard Smith, 1999; McEvoy, 2009; Pacheco et al., 2014). Indeed, the smaller and larger agents’ traversal of their various spaces provides a way to think about how smaller agents’ (cell-level) simple homeostatic loops can scale up into large, organ-level anatomical homeostatic loops. Prentner recently showed how agents build up spatial models of their worlds by taking actions that nullify changes in their experience (Prentner, 2019). Working to nullify changes to one’s state that would otherwise be induced by the vagaries of external environment (and other agents) is the core of homeostasis—the action loops that seek to preserve important states against intervention and entropy. This is not only for physical movement (which results in a creature perceiving itself to be situated in spacetime) but also for other states in which actuation takes place via turning on/off specific genes, remodeling an anatomy, or opening/closing ion channels to change physiological state. An agent can notice patterns in what actions it had to take to keep in homeostasis despite various perturbations that occur, and based on that refine an internal model of some space within which it is acting. This is closely related to the surprise minimization framework (Friston, 2013; Friston et al., 2013; Friston K. et al., 2014), and suggests a straightforward sense in which larger Selves scale up to models of their world and themselves from evolutionary primitives such as metabolic homeostasis. Bioelectricity provides examples where cell-level physiological homeostats form networks that implement much larger-scale pattern memories as attractors, akin to Hopfield networks (Figure 10; Hopfield, 1982; Inoue, 2008; Pietak and Levin, 2017; Cervera et al., 2018, 2019a,b,2020a). This enables all tissues to participate in the kind of pattern completion seen in neural networks—a critical capability for regenerative and developmental repair (anatomical homeostasis). FIGURE 10. FIGURE 10 Open in a new tab Gap junctions scale homeostatic goals. Gap junctions are a type of connection architecture that facilitates scaling of goal states (and thus expands the cognitive cone of cellular agents). A single cell’s homeostatic cycle has 3 parts: measurements, comparison to a stored setpoint, and acting via effectors to stay in or reach the correct region of state space (A). When coupled into gap junctions (B), each of these 3 components expands in size and complexity: the cell group is able to (1) measure a larger region (reacting to spatially more complex inputs, not just local conditions), (2) store a more complex setpoint pattern, and (3) act (deform, grow, etc.) at a scale that produces large-scale anatomical change. The goals of such networks readily map on to regeneration and regulative development (C): dynamical systems pictures of artificial neural networks as they perform pattern completion based on partial input illustrate an energy landscape with wells corresponding to stable target morphology memories. The process of completing a correct planarian pattern from a simple fragment can be modeled in this way, perhaps with overall stress levels instantiating the free energy that the system is trying to minimize (Kuchling et al., 2020b). Such attractors correspond to different possible morphologies, and indeed the normally robust regeneration toward a single pattern (D) can be modified in planaria by temporarily disrupting their gap junctional network, which causes genetically un-modified worms to nevertheless build heads appropriate to other species’ attractors in morphospace (E) (Emmons-Bell et al., 2015; Sullivan et al., 2016). Images in panels (A,B) courtesy of Jeremy Guay of Peregrine Creative. Images in panels (C–E) courtesy of Alexis Pietak. With these pieces in place, it is now possible to mechanistically visualize one specific aspect of the progressive scaling that expands the cognitive light cone. Cells with a chemical receptor can engage in predictive coding to manage their sensory experience (Friston K. et al., 2014; Friston K. J. et al., 2014; Thornton, 2017). Similarly, individual cells homeostatically maintain Vmem (cell membrane resting potential voltage) levels. However, cells can electrically couple via gap junctions to create bioelectric networks that work as a kind of virtual governor—coupled oscillators possess emergent dynamics that now maintain large, spatial patterns of Vmem against perturbation with greater stability (Pietak and Levin, 2017, 2018; Cervera et al., 2018, 2019a,b,2020a,b; Pai et al., 2018; Manicka and Levin, 2019a). These spatial patterns serve as instructive pattern memories guiding the activity of a cell collective through anatomical morphospace toward the correct target morphology (Sullivan et al., 2016; Levin, 2021a; Pezzulo et al., 2021). Voltage is especially interesting because each Vmem level in a single cell is a coarse-grained parameter, subsuming many distinct combinations of sodium, potassium, chloride, etc., levels, and many distinct open/closed states of particular ion channel proteins, which all result in the same resting potential. This can be seen as the minimal version of generalization—cells learning to respond to classes of events by transducing not specific ion levels or channel protein activity states but the macrovariable “voltage.” This in turn enables them to repurpose existing responses for novel combinations of stimuli (e.g., familiar depolarization events caused by novel ion dynamics). Moreover, gap junctions propagate voltage states across tissue, allowing cells to respond to events that are not local in nature (larger-scale) and to respond en masse. More generally, this means that the input to any group of cells is produced by the output of groups of cells—sub-networks, which can be complex and highly processed over time (not instantaneous), enabling predictive coding to manage complex states (at a distance in space and time) and not only purely local, immediate sensory data. It also means that the system is extremely modular, readily connecting diverse upstream and downstream events to each other via the universal adapter of bioelectric states. When this is applied to the homeostatic TOTE (test-operate-exit) loop, allowing its measurement, comparison, and action modules to be independently scaled up (across space, time, and complexity metrics), this inherently expands the cognitive light cone of a homeostatic agent to enable progressively more grandiose goals. Crucially, all of the above-mentioned aspects of the role of generic bioelectric networks underlying the scaling of Selves are not only the products of the evolutionary process, but have many functional implications for evolution itself (forming a positive feedback loop in which rising multiscale agency potentiates the evolution of increasingly more complex versions). Evolutionary Aspects Developmental bioelectricity works alongside other modalities such as gene-regulatory networks, biomechanics, and biochemical systems. The TAME framework emphasizes that what makes it special is that it’s not just another micro-mechanism that developmental biologists need to track. First, developmental bioelectrics is a unique computational layer that provides a tractable entrypoint into the informational architecture and content of the collective intelligence of morphogenesis. Second, bioelectric circuits show examples of modularity, memory, spatial integration, and generalization (abstraction over ion channel microstates)—critical aspects of understanding basal origins of cognition. Developmental bioelectricity provides a bridge between the early problem-solving of body anatomy and the more recent complexity of behavioral sophistication via brains. This unification of two disciplines suggests a number of hypotheses about the evolutionary path that pivoted morphogenetic control mechanisms into the cognitive capacities of behavior, and thus sheds light on how Selves arise and expand. Somatic Bioelectrics Reveals the Origin of Complex Cognitive Systems Developmental bioelectrics is an ancient precursor to nervous systems. Analog bioelectrical dynamics generate patterns in homogenous cell sheets and coordinate information that regulates transcription and cell behaviors. Evolution first exploited this to enable cell groups to position the body configuration in developmental morphospace, long before some cells specialized to use very fast, digital spiking as neural networks for control of behavior as movement in 3-dimensional space (Fields et al., 2020). The function of nervous systems as spatial organizers operating on data from the external world (Keijzer et al., 2013) is an adaptation built upon the prior activity of bioelectric circuits in organizing the internal morphology by processing data from the internal milieu. While neural tissues electrically encode spatial information to guide movement (e.g., memory of a maze in a rat brain) by controlling muscles, bioelectric prepatterns guide the behaviors of other cell types, on slower timescales, during development, regeneration, and remodeling toward invariant, robust anatomical configurations. Developmental bioelectricity illustrates clearly the continuous nature of properties thought to be important for cognition, and the lack of a clean line separating brainy creatures from others. On a single-cell level, even defining a “neuron” is not trivial, as most cells possess the bioelectrical machinery and a large percentage of neuronal genes are also expressed in non-neural cells (Bucher and Anderson, 2015), while neural molecular components are found in cytonemes (Huang et al., 2019). Many channel families were likely already present in the most recent unicellular ancestor (Liebeskind et al., 2015). The phylogeny of ion channels is ancient, and the appearance of context-sensitive channels (enabling new kinds of bioelectrical feedback loops) tracks well with the appearance of complex body plans at the emergence of metazoa (Moran et al., 2015), revealing the remarkable evolutionary continuum that leads from membrane excitability in single cells to cognitive functions in advanced organisms, by way of somatic pattern control (Cook et al., 2014). Fascinating work on bacteria has shown that prokaryotes also utilize bioelectric state for proliferation control (Stratford et al., 2019); and, paralleling the developmental data discussed above, bioelectric phenomena in bacteria scale easily from single-cell properties (Kralj et al., 2011) to the emergence of proto-bodies as bacterial biofilms. Bacterial communities use brain-like bioelectric dynamics to organize tissue-level distribution of metabolites and second messenger molecules, illustrating many of the phenomena observed in complex morphogenetic contexts, such as encoding stable information in membrane potential patterns, bistability, and spatial integration (Humphries et al., 2017; Liu et al., 2017; Larkin et al., 2018; Martinez-Corral et al., 2018; Yang et al., 2020). Not only animal lineages, but plants (Baluska and Mancuso, 2012; Volkov et al., 2019; Serre et al., 2021) use bioelectricity, as evolution frequently exploits the fact that bioelectric dynamics are a powerful and convenient medium for the computations needed to solve problems in a variety of spaces not limited to movement in 3D space. Developmental bioelectricity helps explain how free-living cells scaled their cell-level homeostatic pathways to whole body-level anatomical homeostasis (Levin, 2019). It has long been appreciated that evolvability is potentiated by modularity—the ability to trigger complex morphogenetic cascades by a simple “master” trigger that can be re-deployed in various contexts in the body (von Dassow and Munro, 1999). Recent advances reveal that bioelectric states can form very powerful master inducers that initiate self-limiting organogenesis. For example, the action of a single ion channel can induce an eye-specific bioelectric state that creates complete eyes in gut endoderm, spinal cord, and posterior tissues (Pai et al., 2012)—locations where genetic “master regulators” like the Pax6 transcription factor are insufficient in vertebrates (Chow et al., 1999). Likewise, misexpression of a proton pump (or a 1-h ionophore soak) to trigger bioelectric changes in an amputation wound can induce an entire 8-day cascade of building a complete tadpole tail (Adams et al., 2007; Tseng et al., 2010). This is control at the level of organ, not single cells’ fate specification, and does not require the experimenter to provide all of the information needed to build the complex appendage. Thus, bioelectric states serve as effective master regulators that evolution can exploit to make modular, large-scale changes in anatomy. Moreover, because the same Vmem dynamics can be produced by many different ion channel combinations, and because bioelectric states propagate their influence across tissue distance during morphogenesis (Chernet and Levin, 2014; Pai et al., 2020), evolution is free to swap out channels and explore the bioelectrical state space: simple mutations in electrogenic genes can exert very long-range, highly coordinated changes in anatomy. Indeed, the KCNH8 ion channel and a connexin were identified in the transcriptomic analysis of the evolutionary shift between two functionally different morphologies of fin structures in fish (Kang et al., 2015). The evolutionary significance of bioelectric controls can also be seen across lineages, as some viruses evolved to carry ion channel and gap junction (Vinnexin) genes that enable them to hijack bioelectric machinery used by their target cells (Shimbo et al., 1996; Hover et al., 2017). The unique computational capabilities of bioelectric circuits likely enabled the evolution of nervous systems, as specialized adaptations of the ancient ability of all cell networks to process electrical information as pre-neural networks (Keijzer, 2015; Fields et al., 2020). A full understanding of nervous system function must involve not only its genetics and molecular biology but also the higher levels of organization comprising dynamic physiology and computations involved in memory, decision-making, and spatio-temporal integration. The same is true for the rest of the body. For example, the realization that epithelia are the generators of bioelectric information (Robinson and Messerli, 1996) suggests models in which they act like a retina wrapped around a whole embryo (and individual organs) to preprocess electrical signals into larger-scale features and compute contrast information for downstream processing (Grossberg, 1978). The investigation of somatic bioelectric states as primitive “pattern memories” and the expansion of computational science beyond neurons will enrich the understanding of cell biology at multiple scales beyond molecular mechanisms, as is currently only done with respect to the brain (Marr, 1982). Generalizing the deep concepts of multiscale neuroscience beyond neurons (Grossberg, 1978; Pezzulo and Levin, 2015; Manicka and Levin, 2019b) is necessary for a better understanding of the tissue-level decision-making that drives adaptive development and regeneration. Conversely, advances in understanding information processing in a relatively simpler anatomical context will feed back to enrich important questions in brain science, shedding light on fundamental mechanisms by which information-processing agents (cells) work collectively to accomplish unified, complex system-level outcomes. The multi-disciplinary opportunity here is not only to gain insight into the phylogeny of nervous systems and behavior, but to erase the artificial boundaries between scientific disciplines that focus on neurons vs. the rest of the body, with the direct consequence that a more inclusive, gradualist picture emerges of the mechanisms commonly associated with cognitive Selves. Ion channels and gap junctions are the hardware interface to the bioelectric computational layer within living systems. Like a retina for a brain, or a keyboard for a computer, they allow transient signals to serve as inputs to memory and decision-making networks. For any given agent (cell, tissue, etc.), its bioelectrical interface is accessed by a number of potential users. First are the neighboring agents, such as other tissues, which pass on their bioelectric state during cooperative and competitive interactions in morphogenesis. There are also commensal and parasitic microbes, which have evolved to hijack such control systems to manipulate the anatomy of the host—like the naïve bacteria on planaria that can determine head number and visual system structure in flatworm regeneration (Williams et al., 2020). Moreover, the development of pharmacological, genetic, and optogenetic tools now allows human bioengineers to access bioelectrical circuits for the control of growth and form in regenerative medicine and synthetic bioengineering contexts (Adams et al., 2013, 2014, 2016; Chernet et al., 2016; McNamara et al., 2016; Bonzanni et al., 2020). All of these manipulations can serve as catalysts, enabling an evolutionary lineage to more easily travel to regions of option space that might otherwise be separated by an energy barrier that is difficult for standard evolution to reach. In this sense, cognitive properties of developmental mechanisms help us to understand problem-solving on phylogenetic, not just ontogenetic, timescales. We next look at specific ways in which the architecture of multiscale autonomy, especially as implemented by bioelectric network mechanisms, potentiates evolution. Multi-Scale Autonomy Potentiates the Speed of Evolution Deterministic chaos and complexity theory have made it very clear why bottom-up control of even simple systems (e.g., 3-body problem) can be practically impossible. This inverse problem (Lobo et al., 2014)—what control signals would induce the desired change—is not only a problem for human engineers but also for adjacent biological systems such as the microbiome or a fungus that seeks to control the behavior of an ant (Hughes et al., 2016), and most of all, for other parts of a complex system (to help control itself). Evolution tackles this task by exploiting a multiscale competency architecture (MCA), where subunits making up each level of organization are themselves homeostatic agents. It’s built on an extremely powerful design principle: error correction (Fields and Levin, 2017; Frank, 2019a,b). The key aspect of biological modularity is not simply that complex subroutines can be triggered by simple signals, making it easy to recombine modules in novel ways (Schlosser and Wagner, 2004; Gerhart and Kirschner, 2007), but that these modules are also themselves sub-agents exhibiting infotaxis and socialtaxis, and solving problems in their own spaces (Vergassola et al., 2007; Gottlieb et al., 2013; Karpas et al., 2017). When an eye primordium appears in the wrong place (e.g., a tadpole tail), it still forms a correctly patterned, functional organ, manages to get its data to the brain (via spinal cord connection) to enable vision (Figure 6E), and (if somewhere in the head) moves to the correct place during metamorphosis (Vandenberg et al., 2012). When cells are artificially made to be very large and have several times the normal genetic material, morphogenesis adapts to this and still builds an overall correct animal (Fankhauser, 1945a,b). These are goal-directed (in the cybernetic sense) processes because the system can reach a specific target morphology (and functionality) state despite perturbations or changes in local/starting conditions or the basic underlying components. Regeneration is the most familiar example of this, but is just a special case of the broader phenomenon of anatomical homeostasis. Homeostatic loops operating over large-scale anatomical states have several (closely related) key implications for the power and speed of evolution. First, it greatly smoothes the fitness landscape. Consider two types of organisms: one whose subsystems mechanically follow a hardwired (genetically-specified) set of steps (A, passive, or merely structural modularity), and one whose modules optimize a reward function (B, multi-scale competency of modules). Mutations that would be detrimental in A (e.g., because they move the eye out of its optimal position) are neutral in B, because the competency of the morphogenetic subsystems repositions the eye even if it starts out somewhere else. Thus, MCA shields from selection some aspects of mutations’ negative effects (which inevitably are the bulk of random mutations’ consequences). The primary reason for the anatomical homeostasis seen in regulative development and regeneration may be for dealing, not with damage, but with deviations from target morphology induced by mutations. This is certainly true at the scale of tissues during the lifetime of an individual [as in the inverse relationship between regeneration and cancerous defection from large-scale target morphology (Levin, 2021b)], but may be true on evolutionary time scales as well. Second, MCA reduces apparent pleiotropy—the fact that most mutations have multiple effects (Boyle et al., 2017). For example, a change in an important canonical signaling pathway such as Wnt or BMP (Raible and Ragland, 2005) is going to have numerous consequences for an organism. Suppose a mutation appears that moves the mouth off of its optimal position (bad for fitness) but also has some positive (adaptive) feature elsewhere in the body. In creatures of type A, the positive aspects of that mutation would never be seen by selection because the malfunctioning mouth would reduce the overall fitness or kill the individual outright. However, in creatures of type B, the mouth could move to its optimal spot (Vandenberg et al., 2012), enabling selection to independently evaluate the other consequence of that mutation. Creatures possessing MCA could reap the benefit of positive consequences of a mutation while masking its other effects via local adjustments to new changes that reduce the penalties (an important kind of buffering). In effect, evolution doesn’t have to solve the very difficult search problem of “how to improve feature X without touching features Y. Z which already work well,” and reaps massive efficiency (time) savings by not having to wait until the search process stumbles onto a way to directly encode an improvement that is either isolated from other features, or improves them all simultaneously (Wagner et al., 2007; Melo et al., 2016). Third, MCA allows systems not only to solve problems, but also to exploit opportunities. A lineage has the chance to find out what pro-adaptive things a mutation can do, because competency hides the negative consequences. This gives time for new mutations to appear that hardwire the compensatory changes that had to be applied—an analogy to the proposed Baldwin effect (Hogenson, 2001; Downing, 2004; Robinson and Barron, 2017). This enables the opportunity to exploit the possibility space more freely, providing a kind of patience or reduction of the constraint that evolutionary benefits have to be immediate in order to propagate—it effectively reduces the short-sightedness of the evolutionary process. Indeed, multiscale competency is beneficial not only for natural evolution, but also for soft robotics and synthetic bioengineering because it helps cross the sim-to-real gap: models do not have to be 100% realistic to be predictively useful if the component modules can adaptively make up for some degree of deficiency in the controller design (Brooks, 1986). Fourth, the homeostatic setpoint-seeking architecture makes the relationship between genotype and anatomical phenotype more linear (Muller and Schuppert, 2011; Lobo et al., 2014), improving controllability (Liu et al., 2011; Gao et al., 2014; Posfai et al., 2016). By using a top-down control layer to encode the patterns to which competent subunits operate, living systems do not need to solve the difficult inverse problems of what signals to send their subsystems to achieve high-level outcomes. Bioelectric pattern memories (such as the voltage distribution that tells wild-type planarian cells whether to build 1 head or 2) exploit a separation of data from the machine itself, which makes it much easier to make changes. Evolution does not need to find changes at the micro level but can also simply change the information encoded in the setpoints, such as the electric face prepattern (Vandenberg et al., 2012), which allows it to re-use the same exact implementation machinery to build something that can be quite different. The ability to rely on a non-zero IQ for your component modules (thus delegating and offloading complex regulatory chains) is an important affordance (Watson et al., 2010; Friston et al., 2012) for the evolutionary process. It means that the larger system’s evolution is in effect searching an easier, less convoluted control, signaling, or reward space—this massive dimensionality reduction offers the same advantages human engineers have with agents on the right side of the persuadability scale. It is no accident that learning in the brain, and behavioral systems, eventually exapted this same architecture and indeed the exact same bioelectrical machinery to speed up the benefits of evolution. A significant brake on the efficiency of evolution, as on machine learning (indeed, all learning) is credit assignment: which change or action led to the improvement or reward? When a collection of cells known as a “rat” learns to press a level and get a reward, no individual cell has the experience of interacting with a lever and receiving the nutrient. What enables the associative memory in this collective intelligence are the delay lines (nervous system) between the paws and the intestinal lining which provide a kind of patience—a tolerance of the temporal delay between the action and the reward and the ability to link extremely diverse modules on both ends (different kinds of actions can be linked to arbitrary rewards). MCA does the same thing for evolutionary learning (Watson et al., 2014, 2016; Power et al., 2015; Watson and Szathmary, 2016; Kouvaris et al., 2017), making it easier for systems to reap selection rewards for arbitrary moves in genotype space. This effectively raises the IQ of the evolutionary search process. Much as (Figure 5) an agent’s sophistication can be gauged by how expertly and efficiently it navigates an arbitrary search space and its local optima, the traversal of the evolutionary search process can be made less short-sided by homeostatic activity within the physiological layer that sits between genotype and phenotype. There is an adaptation tradeoff between robustness (e.g., morphogenesis to the same pattern despite interventions and changing conditions) and responsiveness to environment (context sensitivity), perhaps similar to the notion of criticality (Beggs, 2008; Hankey, 2015). The plasticity and goal-directedness of modules (as opposed to hardwired patterns) serve to reduce the sim-to-real gap (Kriegman et al., 2020b): because the current environment always offers novel challenges compared to prior experiences which evolution (or human design) uses to prepare responses, the MCA architecture doesn’t take history too seriously, relying on plasticity and problem-solving more than on fine-tuning micromodels of what to do in specific cases. Biology reaps the benefits of both types of strategies by implementing anatomical homeostasis that coarse-grains robustness by making stability applying to large outcomes, such as overall anatomy, not to the microdetails of cell states. The scaling of homeostatic loops makes it possible to achieve both: consistent results and environmental sensitivity. These dynamics apply in various degrees to the numerous nested, adjacent, and overlapping sub-agents that make up any biological system. Cooperation results not from altruistic actions between Selves, but by the expansion of the borders of a single Self via scaling of the homeostatic loops. On this view, cancer cells are not more selfish than tissues—they are all equally selfish, but maintain goals appropriate to smaller scales of Selves. Indeed, even the parts of one normal body don’t perfectly cooperate—this is as true in development (Gawne et al., 2020) as it is in cognitive science (Dorahy et al., 2014; Reinders et al., 2018, 2019). A picture is emerging of how evolution exploits the local competency of modules, competing and cooperating, to scale these subsystems’ sensing, actuation, and setpoint memories to give rise to coherent larger-scale Selves. Overall, the TAME framework addresses functional aspects only, and is compatible with several views on phenomenal consciousness in compound Selves (Chalmers, 1996). However, it does have a few implications for the study of Consciousness. Consciousness While the TAME framework focuses on 3rd-person observable properties, it does make some commitments to ways of thinking about consciousness. Provisionally, I suggest that consciousness also comes in degrees and kinds (is not binary) for the same reasons argued for continuity of cognition: if consciousness is fundamentally embodied, the plasticity and gradual malleability of bodies suggests that it is a strong requirement for proponents of phase transitions to specify what kind of “atomic” (not further divisible) bodily change makes for a qualitative shift in capacity consciousness. Another implication of TAME is that while “embodiment” is critical for consciousness, it is not restricted to physical bodies acting in 3D space, but also includes perception-action systems working in all sorts of spaces. This implies, counter to many people’s intuitions, that systems that operate in morphogenetic, transcriptional, and other spaces should also have some (if very minimal) degree of consciousness. This in turn suggests that an agent, such as a typical modern human, is really a patchwork of many diverse consciousnesses, only one of which is usually capable of verbally reporting its states (and, not surprisingly, given its limited access and self-boundary, believes itself to be a unitary, sole owner of the body). What is necessary for consciousness? TAME’s perspective is fundamentally that of the primacy of goal-directed activity. Thus, consciousness accompanies specific types of cognitive processes which exert energy toward goals, but as described above, those processes can take forms very divergent from our typical brain-centered view. Unlike other panpsychist views, TAME does not claim that mind is inevitably baked in regardless of physical implementation or structure. Causal structure and cybernetic properties of the embodiment are key determinants of consciousness capacity. However, as the minimal degree of internal self-determination and goal-directedness is apparently present even in particles (Feynman, 1942; Georgiev and Georgiev, 2002; Ogborn et al., 2006; Kaila and Annila, 2008; Ramstead et al., 2019; Kuchling et al., 2020a), there may be no true “0” on the scale of consciousness in the universe. While simple accretion does not magnify the nano-goal-directed activity and indeterminate action of particles (e.g., rocks are not more conscious, and probably less, than particles in specific contexts), biological organization does amplify it, resulting in scaling up of sentience. Of course, these implications will be unpalatable conclusions for many. It should be kept in mind that TAME is compatible with several different views on consciousness, and does not need to commit to one specific philosophy. It is fundamentally a framework for enabling empirical experiments, and its practical utility remains, regardless of the above speculations. Moreover, I remain skeptical about being able to say anything definitive about consciousness per se (as distinct from correlates of consciousness) from a 3rd-person, objective perspective. Thus, however unappealing the above view may be, I see no way of rigorously showing why any other claim about consciousness and what it requires is demonstrably better. An emphasis on somatic plasticity has additional practical implications, being focused on the functional splitting and joining of agents’ parts. For example, the ancient question of “where does it all come together?” in the brain, with respect to the unified character of consciousness, is one of those pseudo-problems that is dispelled by a framework like TAME that focuses on multi-scale architecture. How big should a place where it all comes together be? If it can be ∼140 mm wide, then the answer is, the whole brain. One could decide that it should be smaller (the human pineal gland is ∼7 mm wide), but then the question is, why not smaller still—given the cellular components of the pineal (or any piece of the brain) and the molecular organelles inside a pineal gland cell, one would always need to ask “but where does it all come together inside there?” of whatever piece of the brain is taken to be the seat of consciousness. The multi-scale nature of biology means that there is no privileged size scale for any homunculus. Another important idea with respect to consciousness is “What is it like to be” a given agent (Nagel, 1974). Sensory augmentation, neural link technologies, and bioengineering produce tractable model systems in novel cognitive architectures, such as 2-headed planaria where the brains are connected by a central nervous system (Figure 7B), to help study the functional aspects of this cognitive re-shuffling. TAME’s focus on the fact that all cognitive architectures are inevitably composites emphasizes that the parts can be rearranged; thus, the Subject of cognition can change “on the fly,” not merely during evolutionary timescales. Thus, the basic question of philosophy of mind—what’s it like to be animal X (Nagel, 1974)—is just a first-order step on a much longer journey. The second-order question is, what’s it like to be a caterpillar, slowly changing into a butterfly as its brain is largely dissolved and reassembled into a different architecture for an animal whose sense organs, effectors, and perhaps overall Umwelt is completely different. All of this raises fascinating issues of first person experience not only in purely biological metamorphoses (such as human patients undergoing stem cell implants into their brains), but also technological hybrids such as brains instrumentized with novel sensory arrays, robotic bodies, software information systems, or brains functionally linked to other brains (Warwick et al., 1998; Demarse et al., 2001; Potter et al., 2003; Bakkum et al., 2007a,b; Tsuda et al., 2009; Cohen-Karni et al., 2012; Giselbrecht et al., 2013; Aaser et al., 2017; Ricotti et al., 2017; Ding et al., 2018; Mehrali et al., 2018; Anderson et al., 2020; Ando and Kanzaki, 2020; Merritt et al., 2020; Orive et al., 2020; Saha et al., 2020; Dong et al., 2021; Li et al., 2021; Pio-Lopez, 2021). The developmental approach to the emergence of consciousness on short, ontogenetic timescales complements the related question on phylogenetic timescales, and is likely to be a key component of mature theories in this field. Most surprisingly, the plasticity and capacity for bioengineering and chimerization (recombination of biological and engineered parts in novel configurations) erases the sharp divide between first person and third person perspectives. This has been a fundamental, discrete distinction ever since Descartes, but the capacity for understanding and creating new combinations shows a continuum even in this basic distinction (Figure 11). The fact that Selves are not monadic means we can share parts with our subject of inquiry. If one has to be a system in order to truly know what it’s like to be that system (1st person perspective), this is now possible, to various degrees, by physically merging one’s cognitive architecture with that of another system. Of course, by strongly coupling to another agent, one doesn’t remain the same and experience the other’s consciousness; instead, a new Self is created that is a composite of the two prior individuals and has composite cognition. This is why consciousness research is distinct in strong degree from other scientific topics. One can observe gauges and instruments for 3rd-person science and remain the same Self (largely; the results of the observation may introduce small alterations in the cognitive structure). However, data on 1st person experiential consciousness cannot be taken in without fundamentally changing the Self (being an effective homunculus by watching the neuroscience data corresponding to the movies inside the heads of other people is impossible for the same reason that there is no homunculus in each of our heads). The study of consciousness, whether done via scientific tools or via the mind’s own capacity to change itself, inevitably alters the Subject. Thus, standard (3rd-person) investigations of this process leave open the ancient question as to whether specific upgrades to cognition induce truly discontinuous jumps in consciousness. The TAME framework is not incompatible with novel discoveries about sharp phase transitions, but it takes the null hypothesis to be continuity, and it remains to be seen whether contrary evidence for truly sharp upgrades in consciousness can be provided. Future, radical brain-computer interfaces in human patients are perhaps one avenue where a subject undergoing such a change can convince themselves, and perhaps others, that a qualitative, not continuous, change in their consciousness had occurred. FIGURE 11. FIGURE 11 Open in a new tab Technology reveals gradualism in Descartes’ cut. The apparent fundamental gulf between first person perspective (what is it like to be a specific Self) and third person perspective (external scientific study of that cognitive system) can be seen to also be a gradual continuum, when modern technology is used to expand heterophenomenology (Dennett, 1991). On the left side of the continuum (A) is a traditional 3-rd person scenario of an agent studying another by measuring its physical states: cognitive states can be inferred but not directly experienced (correlates of consciousness), giving rise to the problem of other minds, and a firm distinction between “you” and “me.” However, sensory substitution and augmentation technology now enables the plugging of various peripherals and sensors directly into the nervous system of subjects, and thus it is possible to connect the output of electrophysiology equipment directly into the subject’s brain (B). For example, if the output of a multielectrode array recording neural activity of subject #1 is connected directly to the brainport device (Danilov and Tyler, 2005) of subject #2 (e.g., a scientist), this allows #1’s mental states to more directly provide input into #2’s sensory stream. This can be made even more direct by fusing portions of two brains directly, during embryogenesis, illustrating that the strength of boundaries between “you” and “me” is variable; this configuration may seem far-fetched, but note that it can be readily produced today in animal model systems, and the only barrier to such configurations is ethical, not empirical or logical (C). It’s critical to note that these fusion experiments are not just aberrant corner cases, because all brains are already fusions of neural modules. Single subjects’ brains consist of two hemispheres which must communicate to give rise to a coherent, centralized perception of “me” despite being made of communicating parts, and can be dissociated by commissurotomy (D). Indeed, beyond the two hemispheres, any brain is a collective of smaller active subunits (E) that must all communicate as a collection of cells (each neuron is part of the neighboring neural cell’s “external environment”). This gradient of diverse connections, whether electronic or biological, between and within brains and brain components can be reproduced or expanded upon to whatever degree necessary by biological or technological fusion among subjects. The technological aspect of TAME is that we must develop frameworks that deal not only with standard embodiments of mind as happened to be produced by the path evolution took through life on Earth, but all logically and empirically possible configurations that could evolve, be designed, or both, whether on Earth or in exobiological contexts. The hierarchical, not monadic, structure of cognitive substrates means that the relationship between the parts of one Self and that between a Self and an object of external study is a continuum, not a discrete natural kind. This suggests a key way that actual Consciousness can be studied—by becoming inherently a participant in the experiment, so as to study it from a first-person perspective. Importantly however, what happens when one fuses cognitive systems with their subject of study is that a new Self appears (a composite cognitive system), showing that the Self can remain invariant while pursuing scientific study of functional cognition and behavior (the left of the spectrum), but essentially must change in order to gain first-hand knowledge of consciousness in other cognitive systems. Images are courtesy of Jeremy Guay of Peregrine Creative. With respect to the question of consciousness per se, as opposed to neural or behavioral correlates of consciousness, we have one major functional tool: general anesthesia. It is remarkable that we can readily induce a state in which all the individual cells are fine and healthy, but the larger Self is simply gone [although, some of the parts can continue to learn during this time (Ghoneim and Block, 1997)]. Interestingly, general anesthetics are gap junction blockers (Wentlandt et al., 2006): consistent with the cognitive scaling example above, shutting down electrical communication among the cells leads to a disappearance of the higher-level computational layer while the cellular network is disrupted. GJ blockers are used to anesthetize living beings ranging across plants, Hydra, and human subjects (Gremiaux et al., 2014). It is amazing that the same Self (with memories and other properties) returns, when the anesthetic is removed. Of course, the Self does not return immediately, as shown by the many hallucinatory (Saniova et al., 2009; Kelz et al., 2019) experiences of people coming out of general anesthesia—it takes some time for the brain to return to the correct global bioelectric state once the network connections are allowed again (meta-stability) (Rabinovich et al., 2008). Interestingly, and in line with the proposed isomorphism between cognition and morphogenesis, gap junction blockade has exactly this effect in regeneration: planaria briefly treated with GJ blocker regenerate heads of other species, but eventually snap out of it and remodel back to their correct target morphology (Emmons-Bell et al., 2015). It is no accident that the same reagents cause drastic changes in the high-level Selves in both behavioral and morphogenetic contexts: evolution uses the same scheme (GJ-mediated bioelectrical networks) to implement both. The epistemic problem of Other Minds has been framed to imply that we cannot directly ever be sure how much or what kind of consciousness exists in any particular system under study. The TAME framework reminds us that this is true even for components of ourselves (like the non-verbal brain hemisphere). Perhaps the confabulation system enables one part of our mind to estimate the agency of other parts (the feelings of consciousness and free will) and develop models useful for prediction and control, applying in effect the empirical criteria for persuadability internally. The ability to develop a “theory of mind” about external agents can readily be turned inward, in a composite Self. Are all cognitive systems conscious? The TAME framework is compatible with several views on the nature of consciousness. However, the evolutionary conservation of mechanisms between brains and their non-neural precursors has an important consequence for the question of where consciousness could be found. To the extent that one believes that mechanisms in the brain enable consciousness, all of the same machinery and many similar functional aspects are found in many other places in the body and in other constructs. TAME emphasizes that there is no principled way to restrict consciousness to “human-like, full-blown sophisticated brains,” which means one has to seriously consider degrees of consciousness in other organs, tissues, and synthetic constructs that have the same features neurons and their networks do (Trewavas and Baluska, 2011; Baluska et al., 2016, 2021; Baluska and Reber, 2019). The fundamental gradualism of this framework suggests that whatever consciousness is, some variant and degree thereof has to be present very widely across autopoietic systems. TAME is definitely incompatible with binary views that cut off consciousness at a particular sharp line and it suggests no obvious way to define cognitive systems that have no consciousness whatsoever. A big open question is whether the continuum of cognition (and consciousness) contains a true “0” or only infinitesimal levels for very modest agents. One is tempted to imagine what properties a truly minimal agent would have to have; not being fully constrained by local forces, and ability to pursue goals, both seem key, and both of these are present to a degree in even single particles (via quantum indeterminacy and least action behavior). The type and degree of scaling (or lack thereof) of these capacities in bulk inorganic matter vs. highly-organized living forms is a fertile area for future development of TAME and will be explored in forthcoming work. Conclusion A More Inclusive Framework for Cognition Regenerating, physiological, and behaving systems use effort (energy) to achieve defined, adaptive outcomes despite novel circumstances and unpredictable perturbations. That is a key invariant for cognition; differences in substrate, scale, or origin story among living systems are not fundamental, and obscure an important way to unify key properties of life: the ability to deploy intelligence for problem-solving in diverse domains. Modern theories of Mind must eventually handle the entire option space for intelligent agents, which not only contains the familiar advanced animals we see on Earth, but can also subsume ones consisting of radically different materials, ones created by synthetic bioengineering or combinations of evolution and rational design in the lab, and ones of exobiological as well as possible terrestrial origins. The advances of engineering confirm and put into practice an idea that was already entailed by evolution: that cognitive traits, like all other traits, evolved from humbler variants, forming a continuum. There are no biologically-valid binary categories in this space. Take the prevalent legal definition of human “adults,” who snap into being at the age of 18; such binary views on cognitive properties are fictitious coarse-grainings useful for our legal system to operate, but no more than that. There is no bright line between “truly cognitive” and “pseudo cognitive” that can ever be drawn between two successive members of an evolutionary lineage. The error of “committing Anthropomorphism” is a pseudo-scientific “folk” notion useful for only the most trivial examples of failure to scale down complex claims proportionally to simpler systems; engineering requires us to determine what level of cognitive model enables the most fruitful prediction and control. Every intelligence is a collective intelligence, and the modular, multi-scale architecture of life means that we are a holobiont in more than just the sense of having a microbiome (Chiu and Gilbert, 2015)—we are all patchworks of overlapping, nested, competing, and cooperating agents that have homeostatic (goal-directed) activity within their self-constructed virtual space at a scale that determines their cognitive sophistication. A highly tractable model system for unconventional cognition, in which these processes and the scaling of Selves can not only be seen but can also be manipulated, is morphogenetic homeostasis. The process of construction and remodeling (toward anatomical features) of cellular collectives shows crucial isomorphism to cognitive aspects of the many-into-one binding like credit assignment, learning, stress reduction, etc. The partial wiping of ownership information on permanent signals makes gap junctional coupling an excellent minimal model system for thinking about biological mechanisms that scale cognition while enabling co-existence of subunits with local goals (multiple levels of overlapping Selves, whose scale and borders are porous and can change during the lifetime of the agent). However, many other substrates can no doubt fulfill the same functions. Next Steps: Conceptual and Empirical Research Programs The TAME framework is conceptually incomplete in important ways. On-going development is proceeding along lines including merging with other frameworks such as Active Inference (Friston, 2013; Badcock et al., 2019; Ramstead et al., 2019), Rosen’s (M,R) and Anticipatory Systems (Rosen, 1973, 1979, 1985; Nasuto and Hayashi, 2016), and recent advances in information theory as applied to individuality and scaling of causal power (Hoel et al., 2013, 2016; Krakauer et al., 2014; Daniels et al., 2016). It will be critical to more rigorously develop the waypoints along the Persuadability Continuum, including understanding of what an “increased capacity” human (or non-human) would be like, in contrast to the “diminished capacity” with which we are well familiar from legal proceedings [the right side of the continuum, corresponding to radically expanded cognitive light cones (Śāntideva Bstan ’dzin rgya m and Comité de traduction Padmakara, 2006)]. The TAME framework suggests numerous practical research directions immediately within reach (some of which are already pursued in our group), including developing biomedically-actionable models of morphogenetic plasticity and robustness as meta-cognitive error correction mechanisms, tissue training paradigms for anatomical and physiological outcomes, exploiting learning properties of pathway models for regenerative medicine (Herrera-Delgado et al., 2018; Biswas et al., 2021), and creation of AI platforms based on multi-scale agency architectures that do not rely on neuromorphic principles. Beyond Basic Science: Up-to-Date Ethics The TAME framework also has implications for ethics in several ways. The current emphasis for ethics is on whether bioengineered constructs (e.g., neural cell organoids) are sufficiently like a human brain or not (Hyun et al., 2020), as a criterion for acceptability. Likewise, existing efforts to extend ethics focus on natural, conventional evolutionary products such as invertebrates (Mikhalevich and Powell, 2020). TAME suggests that this is insufficient, because many different architectures for cognition are possible (and will be realized)—similarity to human brains is too parochial and limiting a marker for entities deserving of protection and other moral considerations. We must develop a new ethics that recognizes the diversity of possible minds and bodies, especially since combinations of biological, engineered, and software systems are, and increasingly will be, developed. What something looks like and how it originated (Levin et al., 2020; Bongard and Levin, 2021) will no longer be a good guide when we are confronted with a myriad of creatures that cannot be comfortably placed within the familiar Earth’s phylogenetic tree. Bioengineering of novel Selves raises our moral responsibility. For eons, humans have been creating and releasing into the world advanced, autonomous intelligences—via pregnancy and birth of other humans. This, in Dennett’s phrase, has been achieved until now via high levels of “competency without comprehension” (Dennett, 2017); however, we are now moving into a phase in which we create beings via comprehension—with rational control over their structure and cognitive capacities, which brings additional responsibility. A new ethical framework will have to be formed without reliance on binary folk notions such as “machine,” “robot,” “evolved,” “designed,” etc., because these categories are now seen to not be crisp natural kinds. Instead, wider approaches (such as Buddhist concern for all sentient beings) may be needed to act ethically with respect to agents that have preferences, goals, concerns, and cognitive capacity in very unfamiliar guises. TAME seeks to break through the biases around contingent properties that drive our estimates of who or what deserves proper treatment, to develop a rational, empirically-based mechanism for recognizing Selves around us. Another aspect of ethics is the discussion of limits on technology. Much of it is often driven by a mindset of making sure we don’t run afoul of the risks of negative uses of specific technologies (e.g., genetically-modified organisms in ecosystems). This is of course critical with respect to the new bioengineering capabilities. However, such discussions often are one-sided, framed as if the status quo was excellent, and our main goal is simply to not make things worse. This is a fundamental error which neglects the opportunity cost of failing to fully exploit the technologies which could drive advances in the control of biology. The status quo is not perfect—society faces numerous problems including disparities of quality of life across the globe, incredible suffering from unsolved medical needs, climate change, etc. It must be kept in mind that along with the need to limit negative consequences of scientific research, there is a moral imperative to advance aspects of research programs that will (for example) enable the cracking of the morphogenetic code to revolutionize regenerative medicine far beyond what genomic editing and stem cell biology can do alone (Levin, 2011). The focus on risk arises from a feeling that we should not “mess with nature,” as if the existing structures (from anatomical order to ecosystems) are ideal, and our fumbling attempts will disrupt their delicate balance. While being very careful with powerful advances, it must also be kept in mind that existing balance (i.e., the homeostatic goals of systems from cells to species in the food web) was not achieved by optimizing happiness or any other quality commensurate with modern values: it is the result of dynamical systems properties shaped by the frozen accidents of the meanderings of the evolutionary process and the harsh process of selection for survival capacity. We have the opportunity to use rational design to do better than the basic mechanisms of evolution allow. Importantly, current technologies are forcing us to confront an existential risk. Swarm robotics, Internet of Things, AI, and similar engineering efforts are going to be creating numerous complex, goal-driven systems made up of competent parts. We currently have no mature science of where the goals of such novel Selves come from. TAME reminds us that it is essential to understand how goals of composite entities arise and how they can be predicted and controlled. To avoid the Skynet scenario (Bostrom, 2015), it is imperative to study the scaling of cognition in diverse substrates, so that we can ensure that the goals of powerful, distributed novel beings align with ours. Given the ability of human subunits to merge into even larger (social) structures, how do we construct higher-order Selves that promote flourishing for all? The multicellularity-cancer dynamic (Figure 9) suggests that tight functional connections that blur cognitive boundaries among subunits is a way to increase cooperation and cognitive capacity. However, simply maximizing loss of identity into massive collectivism is a well-known failure at the social level, always resulting in the same dynamic: the goals of the whole diverge sharply from those of the parts, which become as disposable to the larger social Self as shed skin cells are to us. Thus, the goal of this research program beyond biology is the search for optimal binding policies between subunits, which optimize the tradeoffs needed to maximize individual goals and well-being (preserving freedom or empowerment) while reaping the benefits of a scaled-up Self at the level of groups and entire societies. While the specific binding mechanisms used by evolution are not guaranteed to be the policies we want at the social level, the study of these are critical for jump-starting a rigorous program of research into possible ways of scaling that could have social relevance. These issues have been previously addressed in the context of evolutionary dynamics and game theory (Maynard Smith and Szathmáry, 1995; Michod and Nedelcu, 2003; Van Baalen, 2013), but can be significantly expanded using the TAME framework. In the end, important ethical questions around novel agents made of combinations of hardware, software, evolved, and designed components always come back to the nature of the Self. The coherence of a mind, along with its ability to pursue goal-directed activity, is central to our notions of moral responsibility in the legal sense: diminished capacity, and soon, enhanced capacity, to make choices is a pillar for social structures. Mechanist views of cause and effect in the neuroscience of behavior have been said to erode these classical notions. Rather than reduce Selves (to 0, in some eliminativist approaches), TAME (Levin, 2022) finds novel Selves all around us. We see more agency, not less, when evolution and cell biology are taken seriously (Levin and Dennett, 2020). The cognitive Self is not an illusion; what is an illusion is that there is only one, permanent, privileged Self that has to arise entirely bottom-up through the hill-climbing process of evolution. Our goal, at the biomedical, personal, and social levels should not be to destroy or minimize the Self but to recognize it in all its guises, understand its transitions, and enlarge its cognitive capacity toward the well-being of other Selves. Data Availability Statement The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author. Author Contributions ML developed all the ideas and wrote the entire manuscript. Author Disclaimer The opinions expressed in this publication are those of the author(s) and do not necessarily reflect the views of the John Templeton Foundation. Conflict of Interest The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. Publisher’s Note All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher. Acknowledgments I would like to thank Dora Biro, Joshua Bongard, Avery Caulfield, Anna Ciaunica, Pranab Das, Daniel Dennett, Thomas Doctor, Bill Duane, Christopher Fields, Adam Goldstein, EJ, Aysja Johnson, Jeantine Lunshof, Santosh Manicka, Patrick McMillen, Aniruddh Patel, Giovanni Pezzulo, Andrew Reynolds, Elizaveta Solomonova, Matthew Simms, Richard Watson, Olaf Witkowski, Rafael Yuste, and numerous others from the Levin Lab and the Diverse Intelligences community for helpful conversations and discussions, as well as comments on versions of this manuscript. I would also like to thank the three reviewers of the manuscript for important critiques that led to improvement. This manuscript is dedicated to my mother, Luba Levin, who while not having been a scientist, always modeled a deep understanding of, and care for, the multi-scale agency abundant in the world.

A framework for intelligence and cortical function based on grid cells in the neocortex

How the neocortex works is a mystery. In this paper we propose a novel framework for understanding its function. Grid cells are neurons in the entorhinal cortex that represent the location of an animal in its environment. Recent evidence suggests that grid cell-like neurons may also be present in the neocortex. We propose that grid cells exist throughout the neocortex, in every region and in every cortical column. They define a location-based framework for how the neocortex functions. Whereas grid cells in the entorhinal cortex represent the location of one thing, the body relative to its environment, we propose that cortical grid cells simultaneously represent the location of many things. Cortical columns in somatosensory cortex track the location of tactile features relative to the object being touched and cortical columns in visual cortex track the location of visual features relative to the object being viewed. We propose that mechanisms in the entorhinal cortex and hippocampus that evolved for learning the structure of environments are now used by the neocortex to learn the structure of objects. Having a representation of location in each cortical column suggests mechanisms for how the neocortex represents object compositionality and object behaviors. It leads to the hypothesis that every part of the neocortex learns complete models of objects and that there are many models of each object distributed throughout the neocortex. The similarity of circuitry observed in all cortical regions is strong evidence that even high-level cognitive tasks are learned and represented in a location-based framework. Introduction The human neocortex learns an incredibly complex and detailed model of the world. Each of us can recognize 1000s of objects. We know how these objects appear through vision, touch, and audition, we know how these objects behave and change when we interact with them, and we know their location in the world. The human neocortex also learns models of abstract objects, structures that don’t physically exist or that we cannot directly sense. The circuitry of the neocortex is also complex. Understanding how the complex circuitry of the neocortex learns complex models of the world is one of the primary goals of neuroscience. Vernon Mountcastle was the first to propose that all regions of the neocortex are fundamentally the same. What distinguishes one region from another, he argued, is mostly determined by the inputs to a region and not by differences in intrinsic circuitry and function. He further proposed that a small volume of cortex, a cortical column, is the unit of replication (Mountcastle, 1978). These are compelling ideas, but it has been difficult to identify what a column could do that is sufficient to explain all cognitive abilities. Today, the most common view is that the neocortex processes sensory input in a series of hierarchical steps, extracting more and more complex features until objects are recognized (Fukushima, 1980; Riesenhuber and Poggio, 1999). Although this view explains some aspects of sensory inference, it fails to explain the richness of human behavior, how we learn multi-dimensional models of objects, and how we learn how objects themselves change and behave when we interact with them. It also fails to explain what most of the circuitry of the neocortex is doing. In this paper we propose a new theoretical framework based on location processing that addresses many of these shortcomings. Over the past few decades some of the most exciting advances in neuroscience have been related to “grid cells” and “place cells.” These neurons exist in the hippocampal complex of mammals, a set of regions, which, in humans, is roughly the size and shape of a finger, one on each side of the brain. Grid cells in combination with place cells learn maps of the world (O’Keefe and Dostrovsky, 1971; Hafting et al., 2005; Moser et al., 2008). Grid cells represent the current location of an animal relative to those maps. Modeling work on the hippocampus has demonstrated the power of these neural representations for episodic and spatial memory (Byrne et al., 2007; Hasselmo et al., 2010; Hasselmo, 2012), and navigation (Erdem and Hasselmo, 2014; Bush et al., 2015). There is also evidence that grid cells play a role in more abstract cognitive tasks (Constantinescu et al., 2016; Behrens et al., 2018). Recent experimental evidence suggests that grid cells may also be present in the neocortex. Using fMRI (Doeller et al., 2010; Constantinescu et al., 2016; Julian et al., 2018) have found signatures of grid cell-like firing patterns in prefrontal and parietal areas of the neocortex. Using single cell recording in humans (Jacobs et al., 2013) have found more direct evidence of grid cells in frontal cortex (Long and Zhang, 2018), using multiple tetrode recordings, have reported finding cells exhibiting grid cell, place cell, and conjunctive cell responses in rat S1. Our team has proposed that prediction of sensory input by the neocortex requires a representation of an object-centric location to be present throughout the sensory regions of the neocortex, which is consistent with grid cell-like mechanisms (Hawkins et al., 2017). Here we propose that grid cell-like neurons exist in every column of the neocortex. Whereas grid cells in the medial entorhinal cortex (MEC) primarily represent the location of one thing, the body, we suggest that cortical grid cells simultaneously represent the location of multiple things. Columns in somatosensory cortex that receive input from different parts of the body represent the location of those inputs in the external reference frames of the objects being touched. Similarly, cortical columns in visual cortex that receive input from different patches of the retinas represent the location of visual input in the external reference frames of the objects being viewed. Whereas grid cells and place cells learn models of environments via movement of the body, we propose that cortical grid cells combined with sensory input learn models of objects via movement of the sensors. Although much is known about the receptive field properties of grid cells in MEC and how these cells encode location (Rowland et al., 2016), the underlying mechanisms leading to those properties is not known. Experimental results suggest that grid cells have unique membrane and dendritic properties (Domnisoru et al., 2013; Schmidt-Hieber et al., 2017). There are two leading computational candidates, oscillatory interference models (O’Keefe and Burgess, 2005; Burgess et al., 2007; Giocomo et al., 2007, 2011; Burgess, 2008) and continuous attractor models (Fuhs and Touretzky, 2006; Burak and Fiete, 2009). The framework proposed in this paper assumes that “cortical grid cells” exhibit similar physiological properties as grid cells in MEC, but the framework is not dependent on how those properties arise. Throughout this paper we refer to “cortical columns.” We use this term similarly to Mountcastle, to represent a small area of neocortex that spans all layers in depth and of sufficient lateral extent to capture all cell types and receptive field responses. For this paper, a cortical column is not a physically demarked entity. It is a convenience of nomenclature. We typically think of a column as being about one square millimeter of cortical area, although this size is not critical and could vary by species and region. How Grid Cells Represent Location To understand our proposal, we first review how grid cells in the entorhinal cortex are believed to represent space and location, Figure 1. Although many details of grid cell function remain unknown, general consensus exists on the following principles. A grid cell is a neuron that becomes active at multiple locations in an environment, typically in a grid-like, or tiled, triangular lattice. A “grid cell module” is a set of grid cells that activate with the same lattice spacing and orientation but at shifted locations within an environment (Stensola et al., 2012). As an animal moves, the active grid cells in a grid cell module change to reflect the animal’s updated location. This change occurs even if the animal is in the dark, telling us that grid cells are updated using an internal, or “efference,” copy of motor commands (Hafting et al., 2005; McNaughton et al., 2006; Moser et al., 2008; Kropff et al., 2015). This process, called “path integration,” has the desirable property that regardless of the path of movement, when the animal returns to the same physical location, then the same grid cells in a module will be active. Figure 1 www.frontiersin.org Figure 1. How grid cells represent location. (A) An individual grid cell becomes active at multiple locations (red circles) as an animal moves about an environment (rectangle). The locations of activation form a periodic grid-like lattice. The activation locations are always the same for any particular environment. (B) A grid cell module is a set of grid cells that activate at the same spacing and orientation but at different positions in the environment. The activation locations for two grid cells in a grid cell module are shown (red and blue dots). Every location in an environment will activate one or more grid cells in a module. Because of the periodic activation of grid cells, a single grid cell module cannot represent unique locations. (C) Multiple grid cell modules (two shown, top and bottom) tile the same space at different orientations and/or spacings. (D) Although a single module cannot represent unique locations in an environment, the activity across multiple modules can. This rectangle shows the superimposed firing fields of the two grid cells from C). Note that when the two cells (red and green) fire together, only one location is possible (indicated by arrow). The number of locations that can be represented increases exponentially with the number of modules. Due to tiling, a single grid cell module cannot represent a unique location. To form a representation of a unique location requires looking at the active cells in multiple grid cell modules where each grid cell module differs in its tile spacing and/or orientation relative to the environment, Figures 1C,D. For example, if a single grid cell module can represent twenty different locations before repeating, then 10 grid cell modules can represent approximately 2010 different locations before repeating (Fiete et al., 2008). This method of representing location has several desirable properties: (1) Large representational capacity: The number of locations that can be represented by a set of grid cell modules is large as it scales exponentially with the number of modules. (2) Path integration works from any location: No matter what location the network starts with, path integration will work. This is a form of generalization. The path integration properties have to be learned once for each grid cell module, but then apply to all locations, even those the animal has never been in before. (3) Locations are unique to each environment: Every learned environment is associated with a set of unique locations. Experimental recordings suggest that upon entering a learned environment, entorhinal grid cell modules “anchor” differently (Rowland and Moser, 2014; Marozzi et al., 2015). (The term “anchor” refers to selecting which grid cells in each module should be active at the current location.) This suggests that the current location and all the locations that the animal can move to in that environment will, with high certainty, have representations that are unique to that environment (Fiete et al., 2008; Sreenivasan and Fiete, 2011). Combining these properties, we can now broadly describe how grid cells represent an environment such as a room, Figure 2A. An environment consists of a set of location representations that are related to each other via path integration (i.e., the animal can move between these location representations). Each location representation in the set is unique to that environment and will not appear in any other environment. An environment consists of all the locations that the animal can move among, including locations that have not been visited, but could be visited. Associated with some of the location representations are observable landmarks. Figure 2 www.frontiersin.org Figure 2. Representing objects as location spaces. We propose that the neocortex learns the structure of objects in the same way that the entorhinal cortex and hippocampus learn the structure of environments. (A) Two rooms that a rodent has learned. Because of distinct landmarks (suggested by blue and green rectangles) an animal will perceive these as different rooms. Locations in a room are represented by the activity in a set of grid cell modules in the entorhinal cortex. Three locations are shown for each room (A,B,C and D,E,F). Representations of location are unique to both the location in a room and the room. Therefore, if an animal can determine it is in location A, then it knows what room it is in (Room1) and its location in the room. The locations associated with a room are united via movement and path integration. As an animal moves, the representation of location is updated (red arrows) based on an internal copy of its motor behavior. By exploring a room, the animal learns the features associated with locations in the room. (B) We propose that objects such as a pen or coffee cup are similarly defined by a set of locations (four labeled for the pen and three labeled for the cup). Grid cells in the neocortex represent the location of a sensor patch (for example, tip of finger) in the location space of the object. Locations in an object’s space are unique to the object and the location relative to the object. An object’s space includes locations that can be moved to but don’t necessarily have an associated feature. For example, location W is part of the pen because a finger can move from V to W to T via path integration. By moving and exploring the object, the neocortex learns the features associated with locations of the object. Grid Cells in the Neocortex Now let us consider a patch of neocortex that receives input from the tip of a finger, Figure 2B. Our proposal is that some of the neurons in that patch of cortex represent the location of the fingertip as it explores an object. When the finger moves, these cortical grid cells update their representation of location via a motor efference copy and path integration. Objects, such as a coffee cup, have an associated set of locations, in the same way that environments, such as a room, have an associated set of locations. Associated with some of the object’s locations are observable features. The cortical area receiving input from the fingertip tracks the location of the sensory input from the fingertip in the location space of the object. Through movement and sensation, the fingertip cortical area learns models of objects in the same way that grid cells and place cells learn models of environments. Whereas the entorhinal cortex tracks the location of the body, different areas of the neocortex independently track the location of each movable sensory patch. For example, each area of somatosensory cortex tracks the location of sensory input from its associated body part. These areas operate in parallel and build parallel models of objects. The same basic method applies to vision. Patches of the retina are analogous to patches of skin. Different parts of the retina observe different locations on an object. Each patch of cortex receiving visual input tracks the location of its visual input in the location space of the object being observed. As the eyes move, visual cortical columns sense different locations on an object and learn parallel models of the observed object. We have now covered the most basic aspects of our proposal: (1) Every cortical column has neurons that perform a function similar to grid cells. The activation pattern of these cortical grid cells represents the location of the column’s input relative to an external reference frame. The location representation is updated via a motor efference copy and path integration. (2) Cortical columns learn models of objects in the world similarly to how grid cells and place cells learn models of environments. The models learned by cortical columns consist of a set of location representations that are unique to each object, and where some of the locations have observable features. A Location-Based Framework for Cortical Computation Our proposal suggests that cortical columns are more powerful than previously assumed. By pairing input with a grid cell-derived representation of location, individual columns can learn complex models of structure in the world (see also Lewis et al., 2018). In this section we show how a location-based framework allows neurons to learn the rich models that we know the neocortex is capable of. Object Compositionality Objects are composed of other objects arranged in a particular way. For example, it would be inefficient to learn the morphology of a coffee cup by remembering the sensory sensation at each location on the cup. It is far more efficient to learn the cup as the composition of previously learned objects, such as a cylinder and a handle. Consider a coffee cup with a logo on it, Figure 3A. The logo exists in multiple places in the world and is itself a learned “object.” To represent the cup with the logo we need a way of associating one object, “the logo,” at a relative position to another object, “the cup.” Compositional structure is present in almost all objects in the world, therefore cortical columns must have a neural mechanism that represents a new object as an arrangement of previously-learned objects. How can this functionality be achieved? Figure 3 www.frontiersin.org Figure 3. Representing objects as compositions of other objects. (A) The neocortex can learn an object, such as a “coffee cup with logo,” as a composition of two previously learned objects, “cup” and “logo.” The goal is to represent this relationship efficiently, without any relearning. (B) The cup and the logo each have their own unique set of locations. Three locations are shown in cup space (a, b, c) and three locations are shown in logo space (x, y, z). When the logo is placed on the cup there is a fixed one-to-one mapping between locations in cup space and locations in logo space. This relationship can be represented as a displacement vector between the two spaces (blue arrows). (C) Animals exploring an environment can determine the direction and distance from their current location, a, to a previously visited target location, b, even if they have never taken this path before. Determining the displacement between two locations in the same space (e.g., a to b in C) is equivalent to determining the displacement between two locations in separate spaces (e.g., a to x in B). (D) A method to determine the displacement between two locations. Each grid cell module is paired with a displacement cell module. Cells in a displacement cell module (blue dots) respond to a particular displacement between pairs of grid cells (green dots). Any two pairs of grid cells with the same displacement in physical space will activate the same displacement cell. Displacement cells cannot represent a unique displacement in the same way that grid cells cannot represent a unique location. However, the set of active cells in multiple displacement cell modules (three shown) will represent a unique displacement. Because the set of active grid cells in multiple grid cell modules is unique to objects (cup and logo), the set of active displacement cells will also be unique (to both the cup and logo). Thus, a set of active displacement cells can represent the relative placement of two specific objects (location of logo on cup). We have proposed that each object is associated with a set of locations which are unique to the object and comprise a space around the object. If a finger is touching the coffee cup with the logo, then the cortical grid cells representing the location of the finger can at one moment represent the location of the finger in the space of the coffee cup and at another moment, after re-anchoring, represent the location of the finger in the space of the logo. If the logo is attached to the cup, then there is a fixed, one-to-one, relationship between any point in the space of the logo and the equivalent point in the space of the cup, Figure 3B. The task of representing the logo on the cup can be achieved by creating a “displacement” vector that converts any point in cup space to the equivalent point in logo space. Determining the displacement between two objects is similar to a previously-studied navigation problem, specifically, how an animal knows how to get from point a to point b within an environment, Figure 3C. Mechanisms that solve the navigation problem (determining the displacement between two points in the same space) can also solve the object composition problem (determining the displacement between two points in two different spaces). Displacement Cells Several solutions have been proposed for solving the point-to-point navigation problem using grid cells. One class of solutions detects the difference between two sets of active grid cells across multiple grid cell modules (Bush et al., 2015) and another uses linear look-ahead probes using grid cells for planning and computing trajectories (Erdem and Hasselmo, 2014). We suggest an alternate but related solution. Our proposal also relies on detecting differences between two sets of active grid cells, however, we propose this is done on a grid cell module by grid cell module basis. We refer to these cells as “displacement cells” (see Supplementary Material for a more thorough description). Displacement cells are similar to grid cells in that they can’t on their own represent a unique displacement. (In the Supplementary Material example, a displacement cell that represents a displacement of “two to the right and one up,” would also be active for “five over and four up.”) However, the cell activity in multiple displacement cell modules represents a unique displacement in much the same way as the cell activity in multiple grid cell modules represents a unique location, Figure 3D. Hence, a single displacement vector can represent the logo on the coffee cup at a specific relative position. Note, a displacement vector not only represents the relative position of two objects, it also is unique to the two objects. Complex objects can be represented by a set of displacement vectors which define the components of an object and how they are arranged relative to each other. This is a highly efficient means of representing and storing the structure of objects. This method of representing objects allows for hierarchical composition. For example, the logo on the cup is also composed of sub-objects, such as letters and a graphic. A displacement vector placing the logo on the cup implicitly carries with it all the sub-objects of the logo. The method also allows for recursive structures. For example, the logo could contain a picture of a coffee cup with a logo. Hierarchical and recursive composition are fundamental elements of not only physical objects but language, mathematics, and other manifestations of intelligent thought. The key idea is that the identity and relative position of two previously-learned objects, even complex objects, can be represented efficiently by a single displacement vector. Grid Cells and Displacement Cells Perform Complementary Operations Grid cells and displacement cells perform complementary operations. Grid cells determine a new location based on a current location and a displacement vector (i.e., movement). Displacement cells determine what displacement is required to reach a new location from a current location. Grid cells : ( Location1 + Displacement = > Location 2 ) Displacement cells : ( Location2 − Location1 = > Displacement ) If the two locations are in the same space, then grid cells and displacement cells are useful for navigation. In this case, grid cells predict a new location based on a starting location and a given movement. Displacement cells would represent what movement is needed to get from Location1 to Location2. If the two locations are in different spaces (that is the same physical location relative to two different objects) then grid cells and displacement cells are useful for representing the relative position of two objects. Grid cells convert a location in one object space to the equivalent location in a second object space based on a given displacement. In this case, displacement cells represent the relative position of two objects. We propose that grid cells and displacement cells exist in all cortical columns. They perform two fundamental and complementary operations in a location-based framework of cortical processing. By alternating between representations of locations in a single object space and representations of locations in two different object spaces, the neocortex can use grid cells and displacement cells to learn both the structure of objects and generate behaviors to manipulate those objects. The existence of grid cells in the entorhinal cortex is well-documented. We propose they also exist in all regions of the neocortex. The existence of displacement cells is a prediction introduced in this paper. We propose displacement cells are also present in all regions of the neocortex. Given their complementary role to grid cells, it is possible that displacement cells are also present in the hippocampal complex. Object Behaviors Objects may exhibit behaviors. For example, consider the stapler in Figure 4. The top of the stapler can be lifted and rotated. This action changes the stapler’s morphology but not its identity. We don’t perceive the open and closed stapler as two different objects even though the overall shape has changed. The movement of a part of an object relative to other parts of an object is a “behavior” of the object. The behaviors of an object can be learned, and therefore they must be represented in the neural tissue of cortical columns. We can represent behaviors in a location-based framework, again using displacement vectors. The top half and bottom half of the stapler are two components of the stapler. The relative position of the top and bottom is represented by a displacement vector in the same way as the relative position of the logo and the coffee cup. However, unlike the logo on the coffee cup, the two halves of the stapler can move relative to each other. As the stapler top rotates upward, the displacement of the stapler top to bottom changes. Thus, the rotation of the stapler top is represented by a sequence of displacement vectors. By learning this sequence, the system will have learned this behavior of the object. Figure 4 www.frontiersin.org Figure 4. Representing behaviors of objects. Objects have “behaviors,” they can change their shape and features over time. The neocortex can learn these changes, but how? For example, a stapler has several behaviors, one is rotating the top relative to the base. If the top of the stapler is a component object of the stapler, with its own location space, then its position relative to the stapler base is represented by a displacement vector as illustrated in Figure 3. (The top and base of the stapler are analogous to the logo and the cup. Unlike the logo on the cup, the location of the stapler top relative to the base can change.) The closed position is represented by displacement A and the fully open position is represented by displacement E. As the stapler top hinges from the closed to open position, the displacement vector will continually change. (Five positions, A to E, and corresponding displacement vectors are shown.) To learn this behavior, the neocortex only needs to learn the sequence of displacement vectors as the top rotates. Opening and closing the stapler are different behaviors yet they are composed of the same displacement elements, just in reverse order. These are sometimes referred to as “high-order” sequences. Previously we described a neural mechanism for learning high-order sequences in a layer of neurons (Hawkins and Ahmad, 2016). This mechanism, if applied to the displacement modules, would allow the learning, inference, and recall of complex behavioral sequences of objects. “What” and “Where” Processing Sensory processing occurs in two parallel sets of neocortical regions, often referred to as “what” and “where” pathways. In vision, damage to the “what,” or ventral, pathway is associated with the loss of ability to visually recognize objects whereas damage to the “where,” or dorsal, pathway is associated with the loss of ability to reach for an object even if it has been visually identified. Equivalent “what” and “where” pathways have been observed in other sensory modalities, thus it appears to be general principle of cortical organization (Goodale and Milner, 1992; Ungerleider and Haxby, 1994; Rauschecker, 2015). “What” and “where” cortical regions have similar anatomy and therefore we can assume they operate on similar principles. A location-based framework for cortical function is applicable to both “what” and “where” processing. Briefly, we propose that the primary difference between “what” regions and “where” regions is that in “what” regions cortical grid cells represent locations that are allocentric, in the location space of objects, and in “where” regions cortical grid cells represent locations that are egocentric, in the location space of the body. Figure 5 shows how a displacement vector representing movement could be generated in “what” and “where” regions. The basic operation, common to all, is that a region first attends to one location and then to a second location. The displacement cells will determine the movement vector needed to move from the first location to the second location. In a “what” region, Figure 5C, the two locations are in the space of an object, therefore, the displacement vector will represent the movement needed to move the finger from the first location on the object to the second location on the object. In this example, the “what” region needs to know where the finger is relative to the cup, but it does not need to know where the cup or finger is relative to the body. In a “where” region, Figure 5B, the two locations are in the space of the body, therefore, the displacement vector will represent how to move from one egocentric location to a second egocentric location. The “where” region can perform this calculation not knowing what object may or may not be at the second location. A more detailed discussion of processing in “where” regions is beyond the scope of this paper. We only want to point out that it is possible to understand both “what” and “where” processing using similar mechanisms by assuming different location spaces. Figure 5 www.frontiersin.org Figure 5. Location processing in different areas of the brain. Grid cells and displacement cells (see text) can be applied to different tasks in different areas of the brain. (A) If grid cell modules in the hippocampal complex are anchored by cues in an environment, then grid cell activation patterns will represent locations relative to that environment. Given two locations, a and b, displacement cells will calculate the movement vector needed to move the body from point a to point b. (B) If cortical grid cell modules are anchored relative to the body, then they will represent locations in body space. Given two locations, displacement cells will calculate the movement vector needed to move a body part from its current location to a desired new location relative to the body. (C) If cortical grid cell modules are anchored by cues relative to an object, then they will represent locations in the object’s space. Displacement cells will calculate the movement vector needed to move a limb or sensory organ from its current location to a new location relative to the object. Operations performed in (B,C) are associated with “where” and “what” regions in the neocortex. Rethinking Hierarchy, the Thousand Brains Theory of Intelligence Regions of the neocortex are organized in a hierarchy (Felleman and Van Essen, 1991; Riesenhuber and Poggio, 1999; Markov et al., 2014). It is commonly believed that when sensory input enters the neocortex the first region detects simple features. The output of this region is passed to a second region that combines simple features into more complex features. This process is repeated until, several levels up in the hierarchy, cells respond to complete objects (Figure 6A). This view of the neocortex as a hierarchy of feature extractors also underlies many artificial neural networks (LeCun et al., 2015). Figure 6 www.frontiersin.org Figure 6. Rethinking cortical hierarchy. (A) Commonly held view of cortical hierarchy. Sensory input is processed in a hierarchy of cortical regions. The first region detects simple features. The next region combines simple features into more complex features. This is repeated until a region at the top of the hierarchy forms representations of complete objects. (B) Modified view of cortical hierarchy. Every column in every region learns complete models of objects. (Columns learn complete models by combining sensory input with an object-centric location of that input and integrating over movements of the sensor.). Shown are two sensory hierarchies, one for vision and one for touch, both sensing the same object, a cup. There are multiple models of an object within each region, in different regions within a sensory modality, and in different sensory modalities. Although there are many models of the same object (suggested by the small cup images), the models are not identical, as each model is learned via a different subset of the sensory arrays. The green arrows denote the numerically-large cortical-cortical connections that are not hierarchical in nature. The non-hierarchical connections project within the region of origin, across hierarchical levels, across modalities, and between hemispheres. Typically, many columns will be simultaneously observing the same object. The non-hierarchical connections between columns allow them to rapidly infer the correct object (see text). Although learning objects requires movement of the sensors, inference often occurs without movement due to the non-hierarchical connections. We propose that cortical columns are more powerful than currently believed. Every cortical column learns models of complete objects. They achieve this by combining input with a grid cell-derived location, and then integrating over movements (see Hawkins et al., 2017; Lewis et al., 2018 for details). This suggests a modified interpretation of the cortical hierarchy, where complete models of objects are learned at every hierarchical level, and every region contains multiple models of objects (Figure 6B). Feedforward and feedback projections between regions typically connect to multiple levels of the hierarchy (only one level of connection is shown in Figure 6). For example, the retina projects to thalamic relay cells in LGN, which then project to cortical regions V1, V2, and V4, not just V1. This form of “level skipping” is the rule, not the exception. Therefore, V1 and V2 are both, to some extent, operating on retinal input. The connections from LGN to V2 are more divergent suggesting that V2 is learning models at a different spatial scale than V1. We predict that the spatial scale of cortical grid cells in V2 will similarly be larger than those in V1. The level of convergence of input to a region, paired with the spatial scale of its grid cells, determines the range of object sizes the region can learn. For example, imagine recognizing printed letters of the alphabet. Letters at the smallest discernable size will be recognized in V1 and only V1. The direct input to V2 will lack the feature resolution needed. However, larger printed letters would be recognized in both V1 and V2, and even larger letters may be too large for V1 but recognizable in V2. Hierarchical processing still occurs. All we are proposing is that when a region such as V1 passes information to another region such as V2, it is not passing representations of unclassified features but, if it can, it passes representations of complete objects. This would be difficult to observe empirically if objects are represented by population codes as proposed in Hawkins et al. (2017). Individual neurons would participate in many different object representations and if observed in isolation will appear to represent sensory features, not objects. The number of objects that a cortical column can learn is large but limited (Hawkins et al., 2017). Not every column can learn every object. Analysis of system capacity requires a more thorough understanding of hierarchical flow and is beyond the scope of this paper. There are many cortical-cortical projections that are inconsistent with pure hierarchical processing (Figure 6B, green arrows). For example, there are long range projections between regions in the left and right hemispheres (Clarke and Zaidel, 1994), and there are numerous connections between regions in different sensory modalities, even at the lowest levels of the hierarchy (Schroeder and Foxe, 2005; Driver and Noesselt, 2008; Suter and Shepherd, 2015). These connections may not be hierarchical as their axons terminate on cells located outside of cellular layers associated with feedforward or feedback input. It has been estimated that 40% of all possible region-to-region connections actually exist which is much larger than a pure hierarchy would suggest (Felleman and Van Essen, 1991). What is the purpose of these long-range non-hierarchical connections? In Hawkins et al. (2017) we proposed that cell activity in some layers (e.g., L4 and L6) of a column changes with each new sensation, whereas, cell activity in other layers (e.g., L2/3), representing the observed “object,” are stable over changing input. We showed how long-range associative connections in the “object” layer allow multiple columns to vote on what object they are currently observing. For example, if we see and touch a coffee cup there will be many columns simultaneously observing different parts of the cup. These columns will be in multiple levels of both the visual and somatosensory hierarchies. Every one of these columns has a unique sensory input and a unique location, and therefore, long-range connections between the cells representing location and input do not make sense. However, if the columns are observing the same object, then connections between cells in the object layer allow the columns to rapidly settle on the correct object. Thus, non-hierarchical connections between any two regions, even primary and secondary sensory regions in different sensory modalities, make sense if the two regions often observe the same object at the same time (see Hawkins et al., 2017 for details). One of the classic questions about perception is how does the neocortex fuse different sensory inputs into a unified model of a perceived object. We propose that the neocortex implements a decentralized model of sensor fusion. For example, there is no single model of a coffee cup that includes what a cup feels like and looks like. Instead there are 100s of models of a cup. Each model is based on a unique subset of sensory input within different sensory modalities. There will be multiple models based on visual input and multiple models based on somatosensory input. Each model can infer the cup on its own by observing input over movements of its associated sensors. However, long-range non-hierarchical connections allow the models to rapidly reach a consensus of the identity of the underlying object, often in a single sensation. Just because each region learns complete models of objects does not preclude hierarchical flow. The main idea is that the neocortex has 100s, likely 1000s, of models of each object in the world. The integration of observed features does not just occur at the top of the hierarchy, it occurs in every column at all levels of the hierarchy. We call this “The Thousand Brains Theory of Intelligence.” Discussion Crick (1979) wrote an essay titled, “Thinking about the Brain.” In it he wrote, “In spite of the steady accumulation of detailed knowledge, how the human brain works is still profoundly mysterious.” He posited that over the coming years we would undoubtedly accumulate much more data about the brain, but it may not matter, as “our entire way of thinking about such problems may be incorrect.” He concluded that we lacked a “theoretical framework,” a framework in which we can interpret experimental findings and to which detailed theories can be applied. Nearly 40 years after Crick wrote his essay, his observations are still largely valid. Arguably, the most progress we have made toward establishing a theoretical framework is based on the discovery of place cells and grid cells in the hippocampal complex. These discoveries have suggested a framework for how animals learn maps of environments, and how they navigate through the world using these maps. The success of this framework has led to an explosion of interest in studying the entorhinal cortex and hippocampus. In this paper we are proposing a theoretical framework for understanding the neocortex. Our proposed cortical framework is a derivative of the framework established by grid cells and place cells. Mechanisms that evolved for learning the structure of environments are now applied to learning the structure of objects. Mechanisms that evolved for tracking the location of an animal in its environments are now applied to tracking the location of limbs and sensory organs relative to objects in the world. How far this analogy can be taken is uncertain. Within the circuits formed by the hippocampus, subiculum, and entorhinal cortex are grid cells (Hafting et al., 2005), place cells (O’Keefe and Dostrovsky, 1971; O’Keefe and Burgess, 2005), head direction cells (Taube et al., 1990; Giocomo et al., 2014; Winter et al., 2015), border cells (Lever et al., 2009), object vector cells (Deshmukh and Knierim, 2013), and others, plus many conjunctive cells that exhibit properties that are combinations of these (Sargolini et al., 2006; Brandon et al., 2011; Stensola et al., 2012; Hardcastle et al., 2017). We are currently exploring the idea that the neocortex contains cells that perform equivalent functions to the variety of cells found in the hippocampal complex. The properties of these cells would only be detectable in an awake animal actively sensing learned objects. The recent work of Long and Zhang (2018) suggests this might be true. Orientation In the entorhinal cortex, and elsewhere in the brain, are found head direction cells (Taube et al., 1990; Sargolini et al., 2006; Brandon et al., 2011; Giocomo et al., 2014; Winter et al., 2015; Raudies et al., 2016). These cells represent the allocentric orientation of an animal relative to its environment. Inferring where you are via sensation, predicting what you will sense after moving, and determining how to move to get to a new location all require knowing your current orientation relative to your environment. In the models reviewed in Hasselmo (2009) and Hasselmo et al. (2010) head direction cells are critical for accurately transitioning between spatial locations. The same need for orientation exists throughout the neocortex. For example, knowing that a finger is at a particular location on a coffee cup is not sufficient. The finger also has an orientation relative to the cup (which way it is rotated and its angle at contact). Predicting what the finger will sense when it contacts the cup or what movement is required to reach a new location on the cup requires knowing the finger’s orientation relative to the cup in addition to its location. Therefore, we predict that within each cortical column there will be a representation of orientation that performs an analogous function to head direction cells in the hippocampal complex. How orientation is represented in the cortex is unknown. There could be a set of orientation cells each with a preferred orientation, similar to head direction cells, but we are not aware of any evidence for this. Alternately, orientation could be represented via a population code, which would be more difficult to detect. For example, in somatosensory regions orientation could be represented by activating a sparse subset of egocentric orientation detectors (Hsiao et al., 2002; Bensmaia et al., 2008; Pruszynski and Johansson, 2014). How orientation is represented and interacts with cortical grid cells and displacement cells is largely unknown. It is an area we are actively studying. Prediction A long standing principle behind many theories of cortical function is prediction (Lashley, 1951; Rao and Ballard, 1999; Hawkins and Blakeslee, 2004; Lotter et al., 2018). By representing the location of a sensor, a cortical column can associate sensory information within the location space of each object, similar to the way place cells associate sensory information with locations (O’Keefe and Nadel, 1978; Komorowski et al., 2009). This enables a column to build powerful predictive models. For example, when moving your finger from the bottom of a cup to the top, it can predict the sensation regardless of how the cup is rotated with respect to the sensor. Representing composite objects using displacement cells enables a column to generalize and predict sensations even when encountering a novel object. For example, suppose we see a cup with a familiar logo (Figure 3A) and that portions of the logo are obscured. Once a column has recognized the logo and the cup, it can make predictions regarding the entire logo in relation to the cup even if that combined object is new. Building such predictive models would be much harder without an explicit representation of location. In previous papers we proposed dendritic mechanisms that could serve as the neural basis for predictive networks (Hawkins and Ahmad, 2016; Hawkins et al., 2017). Overall, prediction underlies much of the framework discussed in this paper. Attention One of the key elements of a location-based framework for cortical processing is the ability of an area of cortex to rapidly switch between object spaces. To learn there is a logo on the coffee cup we need to alternate our attention between the cup and the logo. With each shift of attention, the cortical grid cells re-anchor to the location space of the newly attended object. This shift to a new object space is necessary to represent the displacement between two objects, such as the logo and the cup. It is normal to continuously shift our attention between the objects around us. With each newly attended object the cortical grid cells re-anchor in the space of the new object, and displacement cells represent where the new object is relative to the previously attended object. Changing attention is intimately tied to movement of the sensor, re-anchoring of grid cells, and, as widely believed, feedback signals to the thalamus (Crick, 1984; McAlonan et al., 2006), presumably to select a subset of input for processing. How these elements work together is poorly understood and represents an area for further study. Uniqueness of Location Code Our proposal is based on the idea that a set of grid cell modules can encode a very large number of unique locations. There are some observations that suggest that grid cells, on their own, may not be capable of forming enough unique codes. For example, because each grid cell exhibits activity over a fairly large area of physical space (Hafting et al., 2005), the activation of the cells in a grid cell module is not very sparse. Sparsity is helpful for creating easily discernable unique codes. The lack of sparsity can be overcome by sampling the activity over more grid cell modules, but not enough is known about the size of grid cell modules and how many can be realistically sampled (Gu et al., 2018) have shown that grid cell modules are composed of smaller sub-units that activate independently, which would also increase the representation capacity of grid cells. Another factor impacting capacity is conjunctive cells. In the entorhinal cortex there are more conjunctive cells than pure grid cells. Conjunctive cells exhibit some combination of “gridness” plus orientation and/or other factors (Sargolini et al., 2006). Conjunctive cells may have a sparser activation than pure grid cells and therefore would be a better basis for forming a set of unique location codes. If the neocortex has cells similar to conjunctive cells, they also might play a role in location coding. Not enough is known about how grid cells, orientation cells, and conjunctive cells work together to suggest exactly how locations are encoded in the neocortex. As we learn more about location coding in the neocortex, it is important to keep these possibilities in mind. Where Are Grid Cells and Displacement Cells in the Neocortex? The neocortex is commonly divided into six layers that run parallel to the surface. There are dozens of different cell types, therefore, each layer contains multiple cell types. Several lines of evidence suggest that cortical grid cells are located in L6 [specifically L6 cortical-cortical neurons (Thomson, 2010)] and displacement cells are located in L5 (specifically L5 thick-tufted neurons) (Figure 7). Figure 7 www.frontiersin.org Figure 7. Location of grid cells and displacement cells in the neocortex. The neocortex contains dozens of cell types commonly organized into six cellular layers. Here, we show a simple drawing of a cortical column. We propose cortical grid cells are located in layer 6 and displacement cells are in layer 5. A requirement of our proposal is that cortical grid cells make bi-lateral connections with displacement cells (solid blue line). Another requirement is that, when combined with a representation of orientation, they make bi-lateral connections with cells in layer 4 (dashed blue line). This is how the column predicts the next input into layer 4. Displacement cells match the unusual connectivity of layer 5 “thick tufted” neurons, which are the motor output cells of the neocortex. These neurons send their axon down into the white matter where the axon splits (green arrows). One branch terminates in sub-cortical structures responsible for motor behavior. The second axon branch terminates on relay cells in the thalamus which become the feedforward input to a hierarchically-higher cortical region. As explained in the text, displacement cells can alternate between representing movements and representing the composition of multiple objects. We propose that L5 thick tufted cells alternate between these two functions which aligns with their unusual connectivity. One piece of evidence suggesting cortical grid cells are in L6 is the unusual connectivity between L4 and L6. L4 is the primary input layer. However, feed forward input forms less than 10% of the synapses on L4 cells (Ahmed et al., 1994, 1997; Sherman and Guillery, 2013), whereas approximately 45% of the synapses on L4 cells come from L6a cortical-cortical neurons (Ahmed et al., 1994; Binzegger et al., 2004). Similarly, L4 cells make large numbers of synapses onto those same L6 cells (McGuire et al., 1984; Binzegger et al., 2004; Kim et al., 2014). Also, the connections between L6 and L4 are relatively narrow in spread (Binzegger et al., 2004). The narrow connectivity between L6 and L4 is reminiscent of the topologically-aligned bidirectional connectivity between grid cells in MEC and place cells in hippocampus (Rowland et al., 2013; Zhang et al., 2013). We previously showed how the reciprocal connections between L6 and L4 can learn the structure of objects by movement of sensors if L6 represents a location in the space of the object (Lewis et al., 2018). For a column to learn the structure of objects in this fashion requires bidirectional connections between cells receiving sensory input and cells representing location. L6a is the only known set of cells that meet this requirement. Also, grid cells use motor input to update their representations for path integration. Experiments show significant motor projections to L6 (Nelson et al., 2013; Leinweber et al., 2017). The current experimental evidence for the presence of grid cells in the neocortex is unfortunately mute on what cortical layers contain grid cells. It should be possible to experimentally determine this in the near future. Our prediction is they will be in L6. The main evidence for displacement cells being in L5 is again connectivity. A subset of L5 cells (known as “L5 thick-tufted cells”) that, as far as we know exists in all cortical regions, projects sub-cortically to brain regions involved with motor behavior. (For example, L5 cells in the visual cortex project to the superior colliculus which controls eye movements.) These L5 cells are the motor output cells of the neocortex. However, the same L5 cells send a branch of their axon to thalamic relay nuclei, which then project to hierarchically higher cortical regions (Douglas and Martin, 2004; Guillery and Sherman, 2011; Sherman and Guillery, 2011). It is difficult to understand how the same L5 cells can be both the motor output and the feedforward input to other regions. One interpretation put forth by Guillery and Sherman is that L5 cells represent a motor command and that the feedforward L5 projection can be interpreted as an efference copy of the motor command (Guillery and Sherman, 2002, 2011). We offer a possible alternate interpretation. The L5 cells in question are displacement cells and they alternately represent movements (sent sub-cortically) and then represent compositional objects (sent to higher regions via thalamic relay cells). As described above, displacement cells will represent a movement vector when comparing two locations in the same space and will represent composite objects when comparing two locations in two different spaces. These two rapidly-changing representations could be disambiguated at their destination either by phase of an oscillatory cycle or by physiological firing patterns (Burgess et al., 2007; Hasselmo, 2008; Hasselmo and Brandon, 2012). Although we are far from having a complete understanding of what the different cellular layers do and how they work together, a location-based framework offers the opportunity of looking anew at the vast body of literature on cortical anatomy and physiology and making progress on this problem. Location-Based Framework for High-Level Thought and Intelligence We have described our location-based framework using examples from sensory inference. Given that the anatomy in all cortical regions is remarkably similar, it is highly likely that everything the neocortex does, including language and other forms of high-level thought, will be built upon the same location-based framework. In support of this idea, the current empirical evidence that grid cells exist in the neocortex was collected from humans performing what might be called “cognitive tasks,” and it was detected in cortical regions that are far from direct sensory input (Doeller et al., 2010; Jacobs et al., 2013; Constantinescu et al., 2016). The location-based framework can be applied to physical structures, such as a cup, and to abstract concepts, such as mathematics and language. A cortical column is fundamentally a system for learning predictive models. The models are learned from inputs and movements that lead to changes in the input. Successful models are ones that can predict the next input given the current state and an anticipated movement. However, the “inputs” and “movements” of a cortical column do not have to correspond to physical entities. The “input” to a column can originate from the retina or it can originate from other regions of the neocortex that have already recognized a visual object such as a word or a mathematical expression. A “movement” can represent the movement of the eyes or it can represent an abstract movement, such as a verb or a mathematical operator. Success in learning a predictive model requires discovering the correct dimensionality of the space of the object, learning how movements update locations in that space, and associating input features with specific locations in the space of the object. These attributes apply to both sensory perception and high-level thought. Imagine a column trying to learn a model of a cup using visual input from the retina and movement input from a finger. This would fail, as the location spaced traversed by the finger would not map onto the feature space of the object as evidenced by the changing inputs from the eyes. Similarly, when trying to understand a mathematical problem you might fail when using one operator to manipulate an equation but succeed by switching to a different operator. Grid cells in the neocortex suggests that all knowledge is learned and stored in the context of locations and location spaces and that “thinking” is movement through those location spaces. We have a long way to go before we understand the details of how the neocortex performs cognitive functions, however, we believe that the location-based framework will not only be at the core of the solutions to these problems, but will suggest solutions. Conclusion It is sometimes said that neuroscience is “data rich and theory poor.” This notion is especially true for the neocortex. We are not lacking empirical data as much as lacking a theoretical framework that can bridge the gap between the heterogeneous capabilities of perception, cognition, and intelligence and the homogeneous circuitry observed in the neocortex. The closest we have to such a framework today is hierarchical feature extraction, which is widely recognized as insufficient. One approach to developing a theory of neocortical function is to build in-silico models of a cortical column based on detailed anatomical data (Helmstaedter et al., 2007; Markram et al., 2015). This approach starts with anatomy and hopes to discover theoretical principles via simulation of a cortical column. We have used a different method. We start with a detailed function that we know the neocortex performs (such as sensory-motor learning and inference), we deduce neural mechanisms that are needed to perform those functions (such as cells that represent location), and then map those neural mechanisms onto detailed biological data. Based on this method, this paper proposes a new framework for understanding how the neocortex works. We propose that grid cells are present everywhere in the neocortex. Cortical grid cells track the location of inputs to the neocortex in the reference frames of the objects being observed. We propose the existence of a new type of neuron, displacement cells, that complement grid cells, and are similarly present throughout the neocortex. The framework shows how it is possible that a small patch of cortex can represent and learn the morphology of objects, how objects are composed of other objects, and the behaviors of objects. The framework also leads to a new interpretation of how the neocortex works overall. Instead of processing input in a series of feature extraction steps leading to object recognition at the top of the hierarchy, the neocortex consists of 1000s of models operating in parallel as well as hierarchically. Introspection can sometimes reveal basic truths that are missed by more objective experimental techniques. As we go about our day we perceive 1000s of objects, such as trees, printed and spoken words, buildings, and people. Everything is perceived at a location. As we attend to each object we perceive the distance and direction from ourselves to these objects, and we perceive where they are relative to each other. The sense of location and distance is inherent to perception, it occurs without effort or delay. It is self-evident that the brain must have neural representations for the locations of objects and for the distances between the objects as we attend to them in succession. The novelty of our claim is that these locations and distances are calculated everywhere in the neocortex, they are the principal data types of cortical function, perception, and intelligence."

Agency in biological and artificial systems

Agency in Natural and Artificial Systems Alvaro Moreno* Arantza Etxeberria Department of Logic and Philosophy of Science University of the Basque Country (UPV-EHU) Avenida de Tolosa 70 20018 San Sebastian Spain ylpmobea@sf.ehu.es ylpetaga@sf.ehu.es * also at Center for Astrobiology (INTA/CSIC) Madrid, Spain Abstract We analyze the conditions for agency in natural and artificial systems. In the case of basic (natural) autonomous systems, self-construction and activity in the environment are two aspects of the same organization, the distinction between which is entirely conceptual: their sensorimotor activities are metabolic, realized according to the same principles and through the same material transformations as those typical of internal processes (such as energy transduction). The two aspects begin to be distinguishable in a particular evolutionary trend, related to the size increase of some groups of organisms whose adaptive abilities depend on motility. Here a specialized system develops, which, in the sensorimotor aspect, is decoupled from the metabolic basis, although it remains dependent on it in the self-constructive aspect. This decoupling reveals a complexification of the organization. In the last section of the article this approach to natural agency is used to analyze artificial systems by posing two problems: whether it is possible to artificially build an organization similar to the natural, and whether this notion of agency can be grounded on different organizing principles. Keywords Agency, autonomy, dynamic decoupling, evolution, motility, robots, sensorimotor activity, size 1 Introduction These days most theoretical and experimental work on behavior, perception, and cognition is per- meated with ecological notions according to which meaning or meaningful behavior arises from dynamic interactions between agents and their environments. Sensorimotor activity is embodied— evolutionarily and developmentally conformed to the structure of the environment (mutualism) — and action and perception are recursively interlocked: action provides occasions for perception and perception guides action. This approach, very influential in Alife-inspired robotics as well as in dynamically oriented theories of cognition, brings together important criticisms of the classical view of cognition as representation of the world (cognitivism). This emphasis on sensorimotor activity has permitted some progress in the elaboration of naturalized models of agency. However, it is not clear whether situated activity via sensorimotor loops is sufficient to define agency and to construct autonomous agents. Two important forerunners of ecological theories, Gibson’s ecological psychology [22] and Varela’s theory of enaction [43, 44], base the study of behavior and cognitive phenomena on situated sensorimotor activity. The two approaches stress different aspects, although both try to dissolve (most) mental representations in sensorimotor activities. Gibson proposes a realist version of what a meaningful environment is, which aims to specify the structure of environmental ‘‘information’’ and how it is captured by organic perceptual systems. Varela’s approach is more focused on the agent, as it is grounded in the earlier notion of autopoiesis, a theory according to n 2005 Massachusetts Institute of Technology Artificial Life 11: 161 – 175 (2005) A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems which an autonomous system creates and maintains its identity as a dynamic network of component production that builds a membrane or boundary between the system inside and the dynamics of the environment. An interesting consequence of an approach of Varela’s type is that all interactions with the environment may be ultimately referred to the task of preserving an autonomous being. A consequence of this is that it is the organism who produces a meaningful world, because the ‘‘meaning’’ is relative to a point of view or perspective created by the very task of self-construction while interacting with a given environment. Nevertheless, although the interlocked nature of self- construction and interaction with the environment is an important aspect of Varela’s approach to understanding life and cognition, his work was not extended to levels higher than that of the minimal autopoiesis. And, from this focus on minimal systems,1 life and cognition appear to be properties of the same kind of organization, or even names given to the same kind of process. This is a confusion that can be clarified if the problem is carefully examined at different levels of organization. In this article we suggest that to define agency we need to consider the task of self-construction while interacting with a given environment (that is to say, the relation with the environment can be considered to serve self-construction). Yet the two aspects of this task, autonomy or self- construction and the relation with the environment, may or may not be entangled. The nucleus of our discussion is precisely the analysis of the different situations for these two aspects: In the natural case, different complexity levels are distinguished, and they are compared with the approaches taken to build artificial systems. We explore the origin and evolution of sensorimotor systems in organisms, especially with regard to motility, and use our findings to analyze the possibility of building robots that may be autonomous in the same sense as we say that organisms are. The argument goes as follows. In the first organisms, metabolic self-construction and adaptive interaction with the environment cannot be distinguished, because both are based on the same mechanism; thus saying that a given process has to do with the maintenance of the system (life) or with dealing with the environment (adaptation) is mostly a conceptual distinction. Yet, as the size of organisms with motility increases, the situation changes, first in eukaryotic cells and later in multicellular organisms. In fact, with the appearance of the nervous system, a new subsystem able to manage the sensorimotor coordination emerges. Although this subsystem is, of course, embedded in the metabolic network of biological processes, it is also, in a way, decoupled from it, for its dynamics is, to a certain extent, underdetermined by the dynamics of metabolic processes. At this point, the conceptual distinction between living or biological processes (to keep the system alive, and even reproduce and evolve) and adaptive or cognitive ones (act in the environment or plan behavior in some other way) acquires a structural correlate, and usually people tend to distinguish life and cognition according to it. In the case of robots, the continuity between life and cognition is missing. Their activity in their world is situated, due to their possession of sensorimotor systems carefully designed under ecological principles, but no process corresponding to the metabolic self-maintenance of organisms can be found. Although the exploration of the abilities of (often minimal) sensorimotor systems is rather sophisticated, including considerations of body form, it is not clear that truly autonomous systems can be obtained this way. Thus, the question is if the decoupling of the nervous system makes it possible to emulate forms of cognitive autonomy based on the operational closure at the level of the nervous system, without basic autonomy. Our concern about whether agents can be defined solely in terms of sensorimotor activity can then be reformulated as this: Is the characteristic dependence found in the living domain a consequence of its evolutionary origins? Can it be generalized to all kinds of agents, including the artificial ones? This article looks at the origin and evolution of agency in living systems (Section 2 1 When studying sensorimotor behavior, the autonomy of the system (operational closure) and the relation with the environment (structural coupling) are studied at the level of the nervous system, and although this is considered to be ‘‘embodied,’’ the relation between the basic or metabolic level of self-construction and the higher or cognitive level permitted by the nervous system is left unspecified. More elaborated work on autonomy can be found in [20]. 162 Artificial Life Volume 11, Number 1 – 2 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems characterizes metabolic agency, and Section 3, neural agency) and compares this situation with the mechanisms of agency production in artificial systems (Section 4). Our conclusion is that living organization depends on a special materiality, and it is still not clear whether robots can fully emulate it. 2 Metabolic Agency It is difficult to offer a good definition for agency. For Kauffman [25], an agent is a system doing something on its own behalf. For Emmeche, ‘‘agency is the establishment of a ‘sentient’ or subjective point of view. . .’’ [17, p. 181]. For Smithers, agents are ‘‘systems that can initiate, sustain, and maintain an ongoing and continuous interaction with their environment as an essential part of their normal functioning’’ [41, p. 97], and for Maes, an agent is simply ‘‘a system that tries to fulfill a set of goals in a complex, dynamic environment’’ [27, p. 2]. Similar definitions can be found in the literature; they mostly emphasize these elements: The activity of the system has goals; the system needs the (dynamical) consequences of its actions for its own (dynamical) endurance or maintenance; the system has a particular point of view. In sum, an agent has the capacity to interact with the environment so as to contribute to its own maintenance. An agent’s action is not merely a physical interaction, because its own viability is affected by it. The actions of an agent are functional in the sense that they serve the viability of the system. A big stone in the river keeps water from flowing, and some bacteria in a milk medium ferment it, producing yogurt. Both do something, but we do not call the stone an agent, whereas the bacteria we can; the difference between the two cases is not in the performed change, but in that the performers gain self-maintenance as a consequence of their actions (the most elemental sense in which we can say that something experiences a change). Thus, one could say that these interactions with the environment are, in an extended sense, a constitutive aspect of the very processes of system organization. If they are oriented, that is to say, if they seek a goal or obey a norm, this will be the maintenance of the acting system itself. Therefore, there is a reciprocal dependence between what defines the subject or the self, and the actions derived from its existence: The being of the system is not separable from its doing. Thus the foundation of agency is the appearance of systems whose organization is maintained because the system actively separates itself from the environment (for example, pumping out ions to conserve a difference in concentration), thus modifying it. 2.1 Formation of Autonomous Systems and Adaptive Capacity Now, where does the agent’s capacity come from? When dealing with a problem of origins we cannot appeal to an external (and certainly, more complex) agent or creator for the system; it is necessary to begin with self-organizing processes in far from equilibrium conditions, namely, with the so-called dissipative structures. Yet, instead of spontaneous dissipative structures, whose maintenance is entirely dependent on a specific set of external boundary conditions, we need to look for the kind of dissipative structure whose organization can be recruited for contributing to its own self-maintenance: Their internal organization should be potentially capable of creating local and selective control mechanisms. However, as Bickhard has pointed out [4], in order to become an agent, a system has to be able not only to perform processes that contribute to its own self-maintenance (a candle flame does that), but also to deploy different processes depending on differences in its environment. In other words, we are looking for a system able to achieve adaptive self-maintenance. Thus, minimal agents are complex dissipative systems that can, under different external conditions, dynamically self- maintain. Now, only chemical systems can develop into complex self-producing dissipative organizations, due to their capacity to form a great variety of plastic recursive component production networks. In organizational terms, recursive chemical systems only become viable autonomous agents when they can build a physical structure that separates them from the external environment, thus creating Artificial Life Volume 11, Number 1 – 2 163 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems specific internal conditions. Maturana and Varela [30] have called such recursive, self-enclosing chemical networks autopoietic systems. An autopoietic system is a component production network that produces its own physical border, and this latter is in turn a necessary constraint for the recursive realization of the former. Now, the realization of an autopoietical system has to solve the following three fundamental problems: (i) the danger of an osmotic crisis, (ii) the spatiotemporal coordination of all processes — and their couplings — and (iii) the achievement of an efficient energy transfer through the components and processes of the system. The minimal set of components necessary to overcome these problems are [38]: a selectively permeable membrane, through which the system controls concentrations, establishes a clear-cut distinction with the environment, and channels the interaction with it; a group of energy currencies (at least one soluble in water, like PPi or ATP, and one directly related to transport processes, like the electrochemical potential gradient of protons or sodium ions across a membrane); and a set of catalysts responsible for modulating the rates at which reactions take place, for setting up regulation (homeostatic) mechanisms, and for carrying out mediated transport processes. The appearance of minimal agents requires the three types of components to be appropriately interacting. Such a system brings about an organizational asymmetry between the inside and the outside domains, separated by a membrane able to select the flow of components across it, via active transport (a hint at the importance of this basic mechanism is that 30% of the whole energy consumed by the cell is used for it). Thus, the system must selectively organize the transport of the surrounding matter and energy for its self-maintenance. Hence, the key element for the origin of agency is functional action as an extension of the recursive self-production processes; active transport is the most basic form of agency, but primitive autonomous systems probably developed other mechanisms of functional action as well (for example, release of chemicals into the environment). The evolutionary path that led from these (hypothetical) autonomous agents to the early living beings about 3.5 billion years ago is unknown. The most transcendental change was probably the introduction of genetic mechanisms to stabilize metabolism [39] — on the one hand, because genetic material allows the production of highly complex proteins (as well as other regulatory factors, as can be seen in [28]), which enhance metabolic plasticity and permit the fine detection of relevant environmental features; and on the other hand, because genetic material makes it possible to store a pool of sequences that can be used to synthesize new proteins in somatic time according to new needs or the detection of new conditions. In present-day bacteria metabolism can make use of genetic information for some (lifetime) adaptive changes, and in certain colonies there is a collective pool of genetic materials, as single individuals horizontally exchange strings via conjugation. 2.2 The Origin of Motility Despite its simplicity, bacterial agency displays many forms of lifetime adaptation: In the case of motility, bacterial organization provides for a wide versatility and complexity of forms (some of them are sessile). Motility is an agent’s capacity to move by its own means, so that it is able to perform fast directional movements to change the environment in looking for preferred conditions. In the world of prokaryotes, both sensors and motility devices are based on very precise proteins, and that is the reason why they have to be understood as adaptive capacities. In bacteria, motility is assisted by membrane devices (flagella), single or multiple, to effectively drive the whole body with rotatory movements (it is interesting that they are the only ‘‘wheel-based’’ motility system in the living domain). Despite their minimal size, bacteria display very efficient forms of motility. Some of them can change their flagella according to the viscosity of the medium: if it is low, they swim across using a polar flagellum, but if it is high, they generate several lateral flagella with which the organism moves with rowing-like movements [3]. Researchers have described a number of bacterial sensors: chemical (chemotaxis), light dependent (phototaxis), magnetic, and so on. The movement of flagellated bacteria (such as E. coli ) consists in a sequence of smooth-swimming runs punctuated by inter- 164 Artificial Life Volume 11, Number 1 – 2 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems mittent tumbles (to randomly change the direction of the runs). They are capable of comparing current and past chemical environments so that the probability of tumbling depends on the chemistry of their immediate surroundings compared with the chemistry they previously encoun- tered. When they run up a gradient of attractants or down a gradient of repellents, they tend to continue their run [47]. Thus, although the bacterium does not detect distant features (it only senses the medium through certain contact proteins), its action as a whole can be interpreted as directed by a distant goal. Another consideration is that what the system can sense is those components that are needed or harmful for its self-maintenance. Yet prokaryote motility is not substantially different from other adaptive capacities, in that we do not find essential organizational differences between different taxa of prokaryotes (although there can be different degrees of complexity). For example, when the prokaryote Caulobacter lives in a very humid medium, it remains fixed to the soil like a vegetal type, whereas in dry periods it reproduces and the new cells grow a flagellum capable of transporting them to a more humid environment. The example shows that movement does not change this organism’s fundamental organization; at this level it is arbitrary to distinguish specifically sensorimotor responses from the metabolic ones, since both are sustained by similar mechanisms. In this sense, the ability to move could be considered as an extension of the set of mechanisms required for self-maintenance. In the smallest living beings, moving to find food (as opposed to capturing primary energy, like sunlight, or fermentation processes) does not require them to develop fundamentally different, or more complex, devices. At this level, all adaptability is based on the maintenance of a chemical organization through the control of energy flows, sometimes aided by very complex proteins, but diffusion is the main way in which components are transmitted inside the cell. Although some investigations involving detailed models of bacterial sensorimotor activity [47], or even proposing artificial models based on them as control mechanisms for robots [48], seem to suggest that there is some capacity for complex information processing, comparatively speaking we find a big difference between this adaptability and other forms that we will discuss in subsequent sections, and it has to do with the existence of specialized devices. 2.3 The Limits of Metabolic Agency At bigger sizes the organization of motility faces new problems raised by the need for fast internal sensorimotor coordination. As eukaryotes are larger than prokaryotes (by about 10,000 times), their surface is smaller relative to their volume: they are bigger inside. That is why, in addition to diffusion processes, eukaryotic cells are equipped with mechanisms to provide precise and speedy internal communication and distribution of substances. Fast adaptive responses are organized in a different manner than in small prokaryotes: they possess microtubules, which contribute not only to chemical channeling and plastic reorganization of selected parts of the internal structure of the cell (cytoskeleton), but also to external movement by means of undulipodia. In eukaryotic cells we find also a functional specialization of subcellular parts (organelles), and some symmetries (anterior- posterior, etc.) appear in their bauplan, which in some cases provide some sense of directionality to the movement. External eukaryotic organs for movement, such as cilia or flagella, have a different, more complex structure than prokaryotic flagella. The construction of motility organs in eukaryotes, based on microtubules, requires larger genes because of the high complexity of the involved proteins (e.g., dynein). Directed motion is a result of the coordinated activity of cilia. In some cases, fast movement is accomplished with the symbiotic collaboration of certain prokaryotes, like spirochaeta, inserted in the membrane of the eukaryotic host. But this new organization conveys a conflict between simultaneous movement and reproduction [9, 29]. According to Buss, a crucial step in the origin of multicellularity is the appearance of gastrulation, in which a hollow ball of cells is transformed into a multi-layered structure with diverse patterns of cell differentiation. His study seeks the origins of multicellularity in the necessity to combine movement and reproduction in single cells. This derives from the observation that the cells Artificial Life Volume 11, Number 1 – 2 165 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems of a metazoan can be either ciliated or prone to divide, but not both. The reason is that both undulipodia (cilia or flagella) and mitotic spindles require microtubule-organizing centers, and in these cells either one structure or the other is possible, but not both. The gastrula is the solution to this problem, where the cells on the surface remain ciliated and those of the interior lose their cilia so they can divide. This conflict is already pointing to an organizational problem that becomes more critical as size increases: the difficulty encountered by the basic metabolic organization in efficiently supporting quick and versatile sensorimotor action. The reason is that as the size grows, it becomes harder for metabolic organization — whose function is the material and energetic self-maintenance of organ- isms — to produce a fast and versatile enough dynamics of internal patterns to coordinate sensor and motor surfaces, allowing quick adaptive movements of the organism. As the size of the organism increases, metabolic organization faces the problem that the energetic and material cost of the production of internal structures for fast sensorimotor coordination becomes incompatible with the accomplishment of the very function of self-maintenance. The path followed by eukaryotes, to grow in size though remaining unicellular, reaches an evolutionary limit with them, and most of the bigger organisms adopted a multicellular organization. Yet, there is a further question that will be very important for our argument: Why do cells not grow any further? Why is the cell size of most organisms constant? According to Bonner, this has to do with energy considerations: ‘‘If one thinks of the rates of different chemical processes occurring within the cell, the distances needed for diffusion, the surface boundaries needed for isolating different chemical components of the motor, and so forth, all of these lead to the conclusion that there is an optimal size with sharp upper and lower limits, which is the size found in nature’’ [7, p. 61]. For others [45], it has to do with the appropriate size to transmit genetic products via diffusion. In any case, large forms of life on Earth have only been able to maintain an efficient metabolism by retaining a cellular organization. 3 Neural Agency An important factor contributing to the appearance of individual multicellular organisms was the big potentiality for cell differentiation exhibited by eukaryotic cells, in particular by those lacking a cell wall. The lack of a cell wall, coupled with the possession of an internal cytoskeleton, permitted the ancestors of animal cells to interact directly with each other through apposed plasma membranes, to adhere to each other, to crawl on surfaces, to differentiate into complex shapes, to engulf other cells by phagocytosis, and to engage in junctional communication with other cells [21, 11]. Thus, new living beings appear, constituted by many functionally differentiated but highly integrated cells. These new multicellular organizations are true organisms, with new forms of agency. At the same time, the significant increase in size of these new multicellular organisms posed further challenges for the organization of motility. Possibly the motion of the first multicellular organisms was simply based on the coordination of local structures: for example, in the case of Porifera, a very primitive type of sessile multicellular organism, responses to stimuli are local and inde- pendent [23]. Yet it is evident that this could not lead to the development of multicellular motility. The problem is that, at this size, it is practically impossible to organize fast and versatile motility with the sole aid of metabolic mechanisms. There are two factors here: the enlarged internal distance between points that need to be connected with small delays (so that the organism can move fast), and the need to selectively modulate the organization of connections (to get adequate sensorimotor correlations). Some multicellular organisms do display fast movement based only on metabolic devices (for example, Dinae plants are able to perform very fast and coordinated movement of millions of cells [40]), but the mechanisms involved lack the flexibility and plasticity required to favor further increase in the complexity of the motor response. Therefore, wherever organisms display fast and versatile movement at a multicellular scale, it is the result of the existence of an internal structure in the organism formed by a type of cell, the neuron, specialized in connecting sensorimotor surfaces 166 Artificial Life Volume 11, Number 1 – 2 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems in a plastic, fast, and (metabolically speaking) cheap way. Neurons differentiated as cells capable of forming branches, which may be interconnected through ion channels in their membranes (controlled either by the electrical potential over the membrane or by ligands). Hence nearly 600 million years ago, the first network of interconnected neurons appears in the Cnidaria, as the most primitive nervous system. 3.1 The Body and the Nervous System In very primitive neural networks there is neither differentiation between sensor, motor, and associative units, nor directionality of nervous impulses, but they are already capable of producing a dynamics of impulse patterns within the interconnected neurons. Unlike chemical signals circulating within the body, which directly interact with metabolic processes, chemical or physical exchanges among neurons allow for recurrent interactions within the network; thus they generate an independent (autonomous) domain of patterns [33]. The dynamics of these patterns is therefore decoupled from the dynamics of metabolic processes, although their material maintenance as well as their functionality still depend upon them. Decoupling means that the organism (or creature) is organized in such a way that (1) part of it (the nervous system) constitutes a relatively independent or autonomous level of interactions whose rules are underdetermined by the dynamics of the remaining system, and (2) the two levels remain causally connected and they depend on each other [31]. Therefore, when we say that the nervous system is a dynamically decoupled system, we mean that nervous interactions obey a dynamics (or a set of rules) not governed by the general metabolic organization, although it is structurally maintained by it. Because of this dynamical decoupling from the metabolic operations, the appearance of the nervous system gave rise to an enormous variety of internal patterns of fast connections, capable of unlimited internal recursivity.2 However, the non-basic, or derived, character of the autonomy of the nervous system has an important implication: The nervous system cannot generate its own functionality by itself. Although the nervous system is a recurrent dynamic network like metabolism, the two differ in that the recursivity of the former is somehow monitored by the basic metabolic level. Neural dynamics only acquire functional meaning if embodied in a self-maintained organization, within a given environ- ment. Configurations become functional when recruited for the tasks of (1) sensorimotor coordination of the animal in its environment, and (2) coordination and fine control of the organism’s metabolic processes. Therefore, the operations of the nervous system ultimately depend on metabolism, on the one hand because its components are cells whose behavior is metabolically organized, and on the other because the nervous system as a whole serves a specific function within the whole organism, which contributes to its capacity to stay alive. Since the global maintenance of the animal’s metabolism requires an adequate sensorimotor activity, nervous system and metabolism are connected in that they depend upon one another for their maintenance (and therefore, existence). Interestingly, this decoupled, but complementary, relation between the neural and the metabolic organizations brings about a body reorganization, which increases the adaptive complexity of multicellular organisms. As a result, individual lifetime interactions do not rely mainly on changes in body structure, but on the neural network. Thus, for these organisms — animals — behavior, understood as functional body movement, is the most important adaptive capacity: their metabolic self-maintenance is largely produced through neurally controlled secretions and motor actions. In addition to quick and efficient motility for organisms with body masses larger than protozoans, the appearance of the nervous system opened up new, qualitatively different modalities of adaptive interaction. Already at early stages of nervous system evolution, rudimentary forms of learning, categorization, and memory appear [1]. We tend to ascribe the main significance of the arrival of 2 One may wonder why large protozoans did not develop some kind of organization similar to the nervous system. Actually, eukaryotic cells already possess complex and diverse subsystems. For example, given the distributed character of the cytoskeleton, it could be considered to be a subsystem of channeled communication capable of producing supple interconnections between membrane sensor and effector structures [24]. However, it is a molecular system; therefore, it cannot be really decoupled from the metabolic organization or endowed with potential for internal recursivity, precisely because it works at the same scale and under the same mechanisms that metabolism does. Artificial Life Volume 11, Number 1 – 2 167 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems nervous systems to these capacities, because they begin to appear as specifically cognitive (insofar as they show increasing similarities with ours). However, the potentialities of the nervous system cannot be developed independently of changes in the general organization of the body. In fact, body organizations and nervous systems of animals coevolve — they shape one another — so that, to a high degree, they are complementary. Not only did the muscular system evolve in close connection with the increase in complexity of the nervous system, but also the organization of internal circulation, the system of fixation, and even the body shape. The appearance and evolution of the nervous system brought along with it changes in the skeletal, respiratory, circulatory, and immune systems, which required building more complex and centralized nervous systems. This fact is manifest when we consider, once again, size increase related to motility: An animal of a certain size requires, in order to be able to move fast, a system that ensures muscle fastening, as well as an adequate flow of nutrients and oxygen to the muscles that execute movement. In the case of a big animal, whose way of life is based on fast movements, the work of its muscles requires a circulatory system, not only closed but also finely regulated in pressure and flow, depending on different internal and environmental circumstances. This is achieved by a system of receptors distributed along the blood vessels, which detect blood pressure, oxygen concentration, and acid level. This also requires a more complex immune system, because the required circulatory system permits easy access of pathogens to any body part. Yet, the very existence of a more complex cardiorespiratory system cannot be conceived without direct control by the nervous system, which regulates cardiac rhythms, arterial pressure, and blood concentrations of certain substances. In vertebrates, for example, a nervous subsystem — the so-called autonomic nervous system — exerts autonomic control of viscera. All this makes manifest that there is a close relation between body and nervous system complexification [31]. Thus, the ability of nervous systems to support complex forms of adaptive behavior is accompanied by new ways of organizing body structures (bauplans). The evolution of agency is to a large extent the history of the interaction between the body and the nervous system: a continuous process of mutual constraining, which sometimes imposes bottlenecks and sometimes enables new evolutionary forms. 3.2 A World of Qualities One of the most dramatic consequences of the development of sensorimotor systems in mobile organisms is that the perceived environment becomes an umwelt, that is to say, a world of qualities affording very sophisticated experiences and activities. In order to be able to see how organisms can make sense of their worlds, we may consider how perceptual systems evolve from very specific ones, each able only to detect one kind of environmental component (in general related to metabolic needs of the system) to more general ones, which not only bring about richer perceptual worlds, but also allow for more complex behaviors or activities. Enzymes, which bind to very specific substrates, are very specific. Many regulating factors affect them in subtle ways and determine the metabolic paths in which components get involved. Nevertheless, each of them recognizes a specific pattern. The sensors that help bacteria swim along (or against) chemical gradients belong to this very class: Protein receptors on the bacterial surface bind with the attractive (or repellent) substance and stimulate the locomotor system [6]. A different case is indirect detection, where there is a deferred relation between the substrate with which the sensor binds and the domain of reality with which the organism interacts. For example, in many organisms vision does not enable a response to light, but to something else. This case can be made clearer by using a borderline example: the primitive acquisition of an ‘‘eye’’ by a paramecium [36, 46]. Some single-celled organisms can use light as nourishment, via photosynthesis, in which light is absorbed directly as energy. They present forms of phototropism: they can sense light and move towards the light. This detection is not usually considered to be vision; it is a case of the mode considered above, of direct detection. Other organisms cannot nourish themselves from light directly, but by eating plants they obtain substances they cannot produce themselves, including vitamins, which enable them to be sensitive to light and to develop eyes. Thus, in evolutionary terms, 168 Artificial Life Volume 11, Number 1 – 2 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems the search for light starts as a search for nourishment, not for information. Wa ¨chterha ¨user reports the case of a paramecium, a single-celled creature, that feeds on green algae and that uses one of them (a chlorella) both as food and literally as eyes [46]. The paramecium sticks the chlorella on itself and maintains a symbiotic relationship with it: as an eye, the chlorella is used to steer the movement of the organism, and as a stomach, it gives back part of its light nourishment to its host. The chlorella couples energetically with the light, but, for the paramecium, the same chlorella is a visual sensor, an eye, used as a means to detect light and thus coordinate its locomotion. This example points to the difference between coupling only with specific substances or specific patterns, and elaborating further the physical, causal effect that certain processes trigger on an organism. The highly evolved perceptual systems of animals can be confusing with regard to what it is to be a sensor or a perceptual system. As a consequence, often perceptual operation is supposed to be completely separated from the energetic/metabolic structure of the organism. This is not the case for those sensors responsible for direct detection. In fact, not all the highly complex perceptual systems share common signal-processing strategies, and this difference is related to the specificity of the information acquired by the organism. A comparison of the visual and olfactory systems of insects shows many differences between them — different neural processing strategies and different neural architectures [34]. Visual perception is influenced by the structure and statistical properties of optical signals, which are complex, but highly constrained. Olfactory signals, on the other hand, carry less information; they are less constrained and predictable. Arbitrary patterns of excitation on the olfactory receptors have to be parsed and learned, and then recognized against a complex background of smells. Vision is hard-wired, its neural circuits tailored to specific behaviors, while olfaction lacks the ordered and highly differentiated neural circuits used for vision and may use an associative network at an early stage to recognize patterns. This difference suggests a further difference in the functions the two systems accomplish. Probably the olfactory system has to recognize specific substances, while the main task of visual elaboration is not to recognize objects, but to elaborate different types of behaviors, for example locomotion. Vision has evolved to allow many different behaviors, some related to locomotion, others to the recognition of a variety of features (relevant for the organism but not always specifically directed to recognition of objects) through an organism-specific elaboration of the properties of one single physical phenomenon: light. Both kinds of perception — direct and indirect — are important for understanding how organisms are involved in meaningful interactions with the environment. Yet one may wonder whether systems with no direct form of perception can autonomously develop indirect ones. 3.3 Towards Minds So far we have sketched the path from basic agency to the arrival and first developments of nervous systems in multicellular organisms. We tried to show that, at the beginning, there is no significant difference between different forms of adaptive capacity: We cannot say that motility per se makes agents more complex, because it is based upon the same metabolic devices used by other strategies (although in the world of prokaryotes there are huge varieties of adaptive forms, some of them more complex than others, all of which rely on the same metabolic principles, occurring at similar speeds). However, a major bifurcation arrives with nervous systems governing multicellular motility, as it implies a qualitative jump in the complexity and variety of agency. When body size increases, agents that adapt via motility require more complex forms of internal organization and interaction. The evolution of multicellulars with nervous systems (animals) brings along a complexification of body structure and organization, as well as of forms of adaptive interaction (behavior, communication), that do not appear in plants or fungi. The emergence of this new form of agency involved wider and more complex interactions with the environment, and it opened a process of increasing complexification of the body organization (an entangled relation between the neural and the basic biological organization) leading to a new, qualitatively different evolutionary process. As we have explained earlier, there are enabling and Artificial Life Volume 11, Number 1 – 2 169 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems disabling constraints that explain why only one special bauplan — the chordates — within the animal kingdom made possible the development of an integrated body and neural changes, able to allow a new increase in body size compatible with versatile movement. In this process, leading to a new complexification of agency, a key feature was the development of new ways of decoupling affecting the nervous system itself, together with different relationships with the rest of the body organization. This new decoupling within the nervous system was crucial for a line of evolution leading to a self-sustained process of encephalization and producing the emergence of new levels of cognitive phenomena, such as emotions and awareness. In the course of the evolution of vertebrates, an increasing part of neural processing is devoted to the internal visceral control tasks and to coordinating all this with sensorimotor activity. At some stage, probably when vertebrates became capable of complex and fast terrestrial movement and their brains attained a certain threshold of complexity, neural activity devoted to the control of behavior through emotions became the basis of what would eventually constitute the mind, understood as a kind of slower, second-order neuroso- matic activity by means of which the animal is able to attain a basic sense of self. This sense of self would be linked to new and more complex forms of coordination between the limbic system and the somatic nervous system [16] and even involving the viscera [13, 14]. Thus, rather than a fast and reactive form of agency, the mind appears to rely on anticipative (non-reactive) control of sensorimotor behavior, based on neural activity, involving an ongoing feedback between the environment, the somatic nervous system, and the autonomous nervous system, mediated by changes in visceral and metabolic states. Probably the non-reactive agency is a consequence of the fact that some neural configurations of the animal, instead of controlling action, contribute to building internal models of reality (virtual interactions with the environment). These internal models, assisted by emotional phenomena, allow for more complex modes of anticipatory behavior, such as self-directed learning, and probably also for some form of awareness [31]. In short, the evolution of agency is linked to the increase in size of those organisms whose way of life is based on motility. Although the evolution of agency took many different paths and contingent modalities, the general principles here stated are probably not contingent. This means that the evolution of any form of organic life in which a tendency to increase in size comes together with adaptive strategies based on motility is likely to develop more complex forms of agency. Only in multicellular organisms whose main form of agency is motility do we find sufficient and necessary conditions for the appearance of cognitive phenomena. Nevertheless we may still question whether those principles are valid outside the natural realm. If what we want is to build physical artificial systems endowed with agential capacities similar to the ones that (natural) organisms exhibit, maybe it is not necessary to base them on the same organizing principles. We will come back to this issue at the end of this article. 4 Embodying Agency: Some Clues for Robotics These days the old dream of reproducing living agency in artificial media appears to many as a feasible one, for, in the last decades, great progress has been made in the manufacture of robots able to perceptually guide their interactions with the environment. However, the extent to which they are autonomous is a matter of perspective. If we compare them with the artificial systems built under the classical AI principles, we have to recognize that they have come a long way from those, and that the field has learned a lot and built many interesting robots. Yet, if we compare them with the organizing principles of living systems explained in previous sections, the situation is very different. Unlike organisms, robots are autonomous in a rather limited sense, more related to situatedness than to self- maintenance. It is evident that their behavior is more their own than it used to be in older devices, but it is not yet self-maintaining or truly autonomous in the same sense as for living beings. Looking at the history of artificial systems emulating living capacities, we observe that some of these capacities seem to be easier to reproduce or emulate than others, and, paradoxically, the more abstract ones (which are intuitively more complex, at least in the sense that it took evolution longer 170 Artificial Life Volume 11, Number 1 – 2 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems to arrive at them) are easier than those that appear simpler (or less evolved). This seeming paradox has to do with the natural inclination human minds possess for abstraction: For us it is a lot easier to abstract than to materialize or embody a given phenomenon. In this context we may pose a further question: whether the creation of autonomous robots has to rely on the organizational principles of organisms at all. We may distinguish two aspects of this question. One is whether it is possible to build an organization similar to that of organisms on a different scale and with different material building blocks. The other is whether it is possible to ground autonomy and cognition on different organizing principles. We will discuss them in the following two subsections. 4.1 Materiality and Organization in Organisms and Robots As we have seen, the autonomy and agency of organisms is based on a particular kind of orga- nization: autopoiesis (self-production) realized via metabolism, that is to say, by a recurrent network of component production. For an organism whose behavior, or whose agency, is continuous with its autopoiesis, the organization is such that actions are a consequence and a cause of the structure of the system. This structure is dissipative: If living beings stop functioning, they will vanish. That is the reason why there is a deep entanglement between what a structure does and what it is. Now, how can such an organization be realized? Can it be achieved through macroscopic components, as a physical, rather than chemical, system? In theory, a recursive organization of component production can be conceived as a physical system, whose building blocks might be microscopic or macroscopic. An example of the latter would be a fully automatized factory, which could reproduce itself starting from simple components. Now, ‘‘fully automatized’’ means that it must be possible to generate all the system devices starting from other devices that are parts of the system. The problem is that, in this case, a great part of the structural complexity of the devices — or even of the building blocks that make up the system — would have to be externally designed, because those building blocks are inert aggregates, which need additional design (i.e., an external level of organization) to become functional constraints harnessing the flows of matter and energy so as to achieve self-construction. For instance, the structure of a microchip does not emerge from its intrinsic composition, but requires external shaping. This system’s organization would not be really self-constructive, because it would depend on a set of external material devices whose design required more complex processes than those produced by themselves, and so on, in an infinite regression. It would never achieve (full) constructive closure, because of the constitutive inertness of the building blocks,3 this organizational logic requires that the external degree of complexity be always greater than the internal one. How is this regression avoided in the case of natural systems? The organization of these systems is really self-produced, because some (macromolecular) components act as devices that constrain the flows of energy, so that the constrained release of energy can be used in turn for generating new macromolecular components acting as constraints, and so on, so that the system can recursively produce the constraints themselves. In other words, the constructive closure of the system is based on a work-constraint cycle [2, 25, 32]. But the realization of such a cycle requires starting with building blocks endowed with certain interactive capacities, derived from their material structure, that is to say, with intrinsically active elements whose combinations may generate new forms of activity. Thus, the system can achieve constructive closure because it creates high-level constraints that act on the (low-level) individual elements, harnessing their dynamics, which in turn recursively produces those control constraints [35]. So, self-constructive autonomy cannot be realized but from a highly complex chemical organization. These chemical components acting as functional constraints should be big molecules; smaller aggregates cannot achieve either the highly specific catalysis nor the storage and replication 3 If building blocks were designed at a molecular scale, they would cease to be inert elements, and in fact we would really be speaking of a chemical system. Artificial Life Volume 11, Number 1 – 2 171 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems of information that are needed (they only yield passively self-organized and self-maintained systems). These components are big enough to act as highly precise devices, but at the same time, small enough to exploit the self-organizing capacities of recurrent chemical networks. Precisely for this reason, in the course of evolution, new, more complex organizations appeared as sets of molecules and intracellular organelles, and later as multicellular structures, still retaining the cellular organiza- tion, as we explained earlier. The conclusion is that living organization is crucially dependent on the materials with which it is built. Hence, to reproduce this organization we would have to use molecular — and even biomolecular — components. 4.2 Functional Autonomy in Organisms and Robots We may now turn to the second question: whether other organizing principles, different from the ones we have described for living systems, can produce systems that are similarly autonomous. In the case of animals, their interaction with the environment is made possible by a sensorimotor system partially decoupled from a basic metabolic organization. What organisms do depends on a complex dynamics emerging from different factors (metabolism, organization of the body, nervous system, condition of the environment, etc.) entangled in a closed organization, whose maintenance is the ultimate goal of behavior. The crucial question is, then, how decoupling can be interpreted in constructive terms, that is to say, whether it suffices to treat the nervous system and its regulation of motility and behavior as a basis for autonomy (defined at an interactive rather than a self-constructive level) or whether we require a dynamic entanglement with a basic infrastructure [41], metabolism (as in organisms). Decoupling implies, at the same time, a dependence of the interactive system on the metabolic one (because it is an integrated system) and the establishment of an independent (or autonomous) dynamics, in the sense that the operation of the higher level system is underdetermined by metabolism. Does this imply that the autonomous behavior of robots can be successfully emulated at this sensorimotor level only, if this independent high-level dynamics is properly grasped? What is the role of metabolism (or other functional systems acting in an integrated organism) in autonomy? In order to answer this question, we need to take into account some characteristics of adaptive or autonomous robots. Their structure and behavior are designed so that there are ongoing couplings between their bodies (at least in biomechanical terms [12]), their nervous systems, and the environment, so that the robot displays real-time adaptive operation: Their behaviors are embodied, embedded, and situated [10, 11, 42]. Thus, rather than being imposed by the programmer, the behavior of these artificial agents arises from real (or realistically simulated) perception-action cycles: they are (or tend to be) able to measure the relevant parameters of the environment to control certain degrees of freedom of the system from the very situation in which the system finds itself (not from the point of view of an external observer), and to act in it physically. This way the results of sensors causally influence motor devices (and vice versa). Recent models even include some kind of functional bodily self-assemblage in the context of the robot’s motor tasks [26], based on the use of some plastic electrochemical building blocks (made of thermoplastic linear activators) that self- assemble according to neural instructions. Hence, these systems explore forms of autonomy at a level that is independent of metabolism, that is to say, of the physical functions responsible for the capacity the creature has to ‘‘stay alive.’’ In self-constructive terms, however, this strategy may produce situatedness, but not autonomy [18]. Although the behavior of the system is controlled by adaptive mechanisms that are inherent in the point of view of the system itself, it is not clear that the system has its own purposes or goals, as it is not the need to maintain its structure that constrains situatedness in its environment. From an approach grounded in the notion of biological organization, situatedness arises as a function of the need to energetically maintain the organism, arising from those processes that endow the agent with identity (i.e., the self-organizing processes that maintain it as a physical system in a physical domain); 172 Artificial Life Volume 11, Number 1 – 2 A. Moreno and A. Etxeberria Agency in Natural and Artificial Systems thus, behavior is a manifestation of the self-maintaining structure, and meaningful interactions emerge from the whole pattern of system-environment relations. In order to emulate that, the actions of the robot should be a consequence and a cause of the structure of the system. One problem is that, in a robot, new functional forms of behavior emerging from its interactions with the environment do not affect its material structure (which is inert), but only its behavioral organization. Thus, the reciprocal causality between what it is and what it does and endures, that exists in the animal, appears in a broken version in the robot [15]. As the robot’s material structure is essentially inert (a quasi-rigid physical structure, mainly conservative), plasticity is obtained at the level of the software, not of the hardware. How can an artificial system of this kind have autonomous goals or purposes? Di Paolo [15] has proposed to address the autonomy of robots at a higher level, so that certain behaviors need not be grounded directly on metabolic needs, nor be dependent on adaptation and survival, but on sets of ‘‘values,’’ or preferences, of the creature emergent at higher levels. It is evident that this approach may provide interesting results for further investigation of the operation of the decoupled high level (which, as we said, is underdetermined by the metabolic one and, in this sense, may even generate values or goals proper to that level), but it does not eliminate the requirement of a strong grounding for the integrated system. In sum, we have tried to show that the functional autonomy of living systems emerges from a structure-process relation that is intrinsically dependent on rates, energy flows, and mass, and that the variations in the values of these elements are crucial for the realization of the closure, and therefore for the organization of agency. This is the reason why in the biological domain the fast-movement – body-size dialectic has brought about the appearance and development of new capacities, such as basic forms of cognition, and (with further development of body plans) an expansion of the cognitive phenomenology and the emergence of a sentient identity (the phenomenon of decoupling played an important role in this evolution). However, in robots the relation between the virtual structure and processes is independent of rate, energy, and matter, and these variations do not substantially affect the organization of agency. As a consequence, we should not expect that work in artificial systems will generate forms of agency similar to that in living ones. 5 Conclusion In this article we have analyzed the conditions for agency in natural and artificial systems. In living beings, agency arose through an evolutionary process initiated by basic forms of autonomy, whereas in the construction of robots and simulated artificial agents the starting point is human cognitive capacities. We have tried to show that this difference in beginnings is concurrent with a difference in principles. In animals the phenomenon of decoupling reveals a complexification of the basic organization, because the sensorimotor system serves a function in the integrated whole organism, whereas its characteristic dynamics is somehow autonomous. So far, the research program in robotics and artificial intelligence has focused on autonomy at the sensorimotor level, without basic autonomy. Instead of trying to reconstruct the conditions for triggering a process of spontaneous organization starting from chemical components, computers and other complex technical tools are the building blocks. The lesson of artificial life is that we can build ‘‘creative’’ artifacts in which human action appears only in an indirect way, ‘‘getting ourselves out of the loop,’’ as Langton says [5, 37]. From this perspective, evolution in the computer may generate functional structures without full design (as is the aim of evolutionary robotics), yet artificial evolution is different form natural evolution, and the two evolutionary trends will have different characteristics [19]. The reason for the (in spite of all progress) still evanescent expectation of reproducing true agency and cognition in robots is not that there is something extra in living matter that we do not know (an idea that brings old vitalisms back to life). It is not that computer power needs to be improved or that the organizing or mathematical principles need to be reformulated, as Brooks Artificial Life Volume 11, Number 1 – 2 173 A.