Evolution "Doesn't Need" Mutation - Blaise Agüera y Arcas
Blaise Agüera y Arcas presents his research on artificial life experiments showing how evolution can create complex programs without mutation through symbiogenesis - the fusion of simpler replicators. He demonstrates that life emerges as a phase transition in computational systems, with complexity arising through cooperation and merger of functional components rather than random mutations.
- Evolution doesn't require mutation to create complexity - symbiogenesis (fusion of replicators) is the primary driver of evolutionary novelty
- Life represents a distinct phase of matter characterized by embodied computation and autopoiesis, emerging through phase transitions
- Intelligence and life are fundamentally connected from the beginning, as living systems must model themselves and their environment
- Genomes are fractal structures of replicators within replicators, with most complexity arising from viral integrations and horizontal transfers
- The arrow of time in evolution comes from symbiogenetic events that necessarily increase informational complexity
"After a few million interactions, magic happens, which is that you go from noise to programs. You start to see complex programs appear on these tapes."
"Life is literally embodied computation. It is computational. You cannot have life without having computation."
"Function is the thing that life has that non life doesn't have."
"Life is an embodied autopoietic computation arising and complexifying through symbiogenesis."
"It's symbiogenesis all the way down. Most of these symbiogenic events are much more uneven."
Ugh. You said you were over him, but his hoodie's still in your rotation. It's time. Grab your phone, snap a few pics and sell it on depop. Listed in minutes with no selling fees. And just like that, a guy 500 miles away just paid full price for your closure. And right on cue. Hey, still got my hoodie? Nope. But I've got tonight's dinner paid for. Start selling on depop, where taste recognizes taste list. Now with no selling fees, payment processing fees and boosting fees still apply. See website for details. We heard you. Nine years of bring back the snack wrap and you've won. But maybe you should have asked for more. Say hello to the hot honey snack wrap. Now you've really won. Go to McDonald's and get it while you can. After a few million interactions, magic happens, which is that you go from noise to programs. You start to see complex programs appear on these tapes. This is the most exciting plot that I've made in the last few years. And it's the one that's on the COVID of the book. You can see that in the beginning it's not very computational. And then a sudden transition takes place here. It looks like a phase transition. This is the book that I hear is making the rounds at Sakana, which I'm very happy to hear. The big one on the right, what is Intelligence? Is sort of the Lord of the Rings and what is Life? On the left is kind of the Hobbit. So it's kind of the single. And it's also chapter one of what is Intelligence? So it goes kind of inside the other one. Mostly what I'll be talking about today is what's in these two books, but with quite a bit more detail, more mathematical detail, since I think this is a really good audience for that. And I'll also be connecting it a bit with some of the bigger themes of the A Life conference and community and dare I say, even movement. In particular, I actually wanted to begin with this wonderful sort of Open Problems in Artificial Life summary paper which has a number of very illustrious co authors, at least one of whom we heard from yesterday and more than one of whom are here at the conference. This is open problems 14 open problems in artificial life in the year 2000. How does life arise from the nonliving? How do the transition to life in an artificial chemistry or an in silica environment can occur and why it occurs? I'm sure many of you know this was the problem that bedeviled Darwin. He made one of the most rich and explanatorily powerful theories ever in science in discovering how evolution works. But he was unable to explain how evolution got started. He at some point in one of his letters said, you might as well talk about the origin of matter. I think that the origin of matter and the origin of life might actually be one and the same thing, and evolution might actually be the answer to that question. But it's an evolution that includes a term that Darwin did not account for in his original formulation. In section B of these questions determine what is inevitable in the open ended evolution of life. I'm hoping to speak a little bit about that too. Create a formal framework for synthesizing dynamical hierarchies at all scales and develop a theory of information processing, information flow and information generation for evolving systems. I won't be going into the information theory in any detail, but hopefully will set up the problem in a perhaps somewhat new way that I hope will help to do that. And finally, in section C, how is life related to mind, machines and culture? If I have time, I will get into this as well and talk a bit about the emergence of intelligence and mind in an artificial living system and the influence of machines on the next major evolutionary transition of life. So it was really cool to read this paper from 2020 and to see how much of the perspective that you had already been exploring then feels right and consistent with a sort of fresh look at these problems in 2025. Let me just begin with this question of souls. It used to be in the 19th century and earlier that we thought that life had some vital force or spirit that animated it and made it different From Inan the 19th century when we began to figure out organic chemistry and be able to synthesize urea and so on. The idea that no, we should really adopt a strictly materialist perspective because there's nothing special or different about the matter in us versus the matter anywhere else in the universe took hold. And that's progress for sure. But it also, when we embrace atoms and materialism fully, we're left with some questions about what differentiates life from non life. Then what can we even say about life? There are at least some biologists who say, well, maybe it's not even meaningful to talk about any difference between life and non life. But I don't think that that's true. And I think that the answer to the conundrum is to invoke function. Function is the thing that life has that non life doesn't have. In other words, just to give you a little parable, if I were to come back from the future with this object and you ask me what it is, and I tell you it is an artificial kidney with 100 year lifespan. You can implant it in a body and it'll work the way your kidneys do. It'll filter urea from the blood and so on. That's a really important piece of information. But it's not a material or a materialist piece of information. It's not something that you could read off from the atoms. And those atoms could be, I don't know, tungsten filaments or carbon nanotubes are made out of some technology we don't understand now. Or it could be organic, it could be made out of cloned tissue. And the point is that it working as a kidney doesn't depend on that matter. There is a kind of separation of concerns between the matter and the function. And so there's some real sense in which the function is like a spirit or like something immaterial. It's not material. And yet it also relies, of course, on the physics of what's going on. You can't have the spirit without the matter, as it were. So function is really important. And function is something that a rock on a non living planet somewhere doesn't have. If you break a rock on a nonliving planet, you now have two rocks. You don't have a broken rock. If you break a kidney, you now no longer have a working kidney. That's the difference between something functional and something non functional. This idea of function was formalized by Alan Turing, who never intended the Turing machine to actually be built when he wrote it in 1936. But there is one that was built by Mike Davy in 2010. I don't need to review Turing machines with all of you, of course, you all know how they work. But I do want to review briefly von Neumann's update to Turing's thinking about computation, which he did a few years later. This was published posthumously after von Neumann died. But the idea behind von Neumann's thinking is he was trying to answer the same question that Schrodinger had asked in his what is Life? Book. And in particular, he was trying to ask the question, if you have a robot that is swimming around in a pond and the pond has lots of loose Legos around, I don't know if there were Legos in 1950, but let's pretend they were Legos in 1950, and the job of the robot is to assemble those Legos into a new robot like itself. There's something a little Bit mysterious about that. It feels a little bit like pulling yourself up by your own bootstraps, or like a paradox. And so he asked, what does it take for something to be able to make something like itself? Which seems hard, almost paradoxical. And his conclusion was, well, you need to have instructions for how to make a mii. You need to have a tape with instructions for how to make a mii, and you need to have a universal constructor that will follow the instructions on that tape in order to assemble the necessary parts. And you also need to have a tape copier so that you can give your offspring a copy of that tape. And by the way, the tape has to also include the instructions for making the universal constructor and the tape copier. And if those things all hold, then you have life, you have something that can build itself. And what's so profound about von Neumann's insight? I mean, first of all, he predicted all of this before we knew the structure and function of DNA, before we understood what ribosomes were or had discovered DNA polymerase, so he called it. Exactly right. All of those things really do exist inside cells. And he figured this out from pure theory, never having set foot in a biolab. The profound insight is that he said, by the way, a universal constructor is a universal Turing machine. Those are literally one and the same thing. And by making that observation, what he discovered was that life is literally embodied computation. It is computational. You cannot have life without having computation. So obviously not everything that is alive reproduces, but everything that is alive has to be able to make itself. It has to be able to do some combination of healing, growing, maintaining itself, reproducing. All of that is autopoiesis, all of that involves self construction, and all of that necessarily involves a universal constructor. Now, what do I mean by embodied computation? This is a really important distinction between von Neumann and Turing. In Turing, the symbols that the head writes are different from the head itself and the tape and the table of rules that the head follows. Whereas in von Neumann, it's more like a 3D printer. The memory is atoms, not abstract symbols. In other words, you could think about a Turing machine as like this laptop, which can't extrude another laptop out the side. But a von Neumann replicator is like a combination of a laptop and a 3D printer that can print another laptop. So its memory is actually atoms. That's what I mean by embodied. So I don't mean embodied in the ways that a lot of roboticists talk about embodied. I mean that there is a closure between the medium in which the computation happens and the thing that is actually doing the computation, that's the key. So computation that is embodied in that sense and that is autopoietic, is alive. You can't reproduce non trivially, evolvably without computation. No computation, no life. I do want to say a word briefly about what I mean by computation. And in this I'm following the work of Susan Stepney, Dominic Horsman, Rob Wagner, Viv Kendon. This is from a nice paper they wrote in 2023 relating the evolution of a physical system and the computation that it does. So on top you have logical gates, on the bottom you have transistors in your computer. This is important because there are no bits in a computer. There are just voltages that go up and down. In fact, even the voltages are an abstraction. Something further, if we go further down. But the point is that you have to coarse grain those voltages into bits and then you have to have a logical machine that talks about how those bits evolve, what are the computational processes that those bits undergo. And there's a mapping from the physical system to the logical system, and vice versa. When we say something computes, what we mean is that it is possible to construct such a mapping and that therefore, as the physical system evolves, that is equivalent to the logical system evolving. So there are some caveats. You can have stochastic computation in which there's a little bit of randomness injected, so it doesn't have to be fully deterministic. Another really important caveat is that you don't want that description to be infinitely complex. Otherwise you could have the trivial case of saying, like the water in the Seine is a computer and the longer my computation, I just need to make my description longer and longer in order to match. No, that doesn't work either. You need a kind of Occam's Razor description for it to be valid. But this is a good definition of computation. But it emphasizes that there is something subjective about computation. You need to have a model for how the physical system translates into the logical system. In order for any of this stuff to work. There are implications about entropy, free energy and heat and so on in this model. And in particular, as you all know, we've talked already, Hector Zanil in his very elegant talk of a couple of days ago, talked about, and actually Chris Kempis also talked about the Landauer limit and the fact that in a computational system you're constantly reducing the entropy of your state space, and in doing so you therefore require free energy so you need to have free energy available and you need to eject waste heat. The exception, in a way, only proves the rule, which is reversible computation. In reversible computation you generate ansyllabates, and that's equivalent to just saying there's no exhaust. But then you either have to keep on making your computer bigger and bigger and bigger as you accumulate these ansyllabates, or you have to shrink what you consider to be the computer, and then you're back to non reversible computation once again. Three important fallacies that I want to point out before continuing. One of them I will call the Sapolsky error. Robert Sapolsky has written famously about people not having because we're built on physical systems, the physics is, if you like, deterministic. Let's set aside quantum mechanics and stuff like this. Let's imagine we live in a Newtonian universe. It's fine, it's good enough. The point is that physics is reversible. All of the basic physics that we understand, whether that's Newton's equations, Maxwell's equations, Einstein's equations, quantum mechanics, all of those are essentially time reversible. So you can move them either forward or back. Computation is not reversible. When I add three plus five to get eight, once I've got the eight and I haven't kept my ansyllabates around, let's say I no longer know what was added in order to make the 8, computation is inherently irreversible. And so to say that what is true of the physical system is also true of the computational system, of the logical system is not the case. And reversibility would be one trivial example of how that is not the case. Causation, by the way, only makes sense in the light of irreversibility. So if you have a purely physical system, then to say that A causes B is equivalent to saying that B causes A. Because everything is kind of a block universe, if you like, in that kind of setup. But in computation you can talk about causality because there are ifs and thens in there. And this once again connects with the way Hector was talking about how essentially nothing in causation makes sense except in the light of computation, which I completely agree with another fallacy we could call the early Wittgenstein error. If we say something like birds exist in the world line one of the tractatos logical philosophy didn't say birds, but whatever, you can't say birds exist or birds don't exist in a way that is independent of A model of the universe. There are no birds in physics. There are no birds in this underlying dynamical system. When we start talking about birds, we already are talking about having some kind of model. And once you start talking about models, you've got causality, reversibility, irreversibility, all kinds of other things in play. And none of these statements are airtight. They all rely on an observer. This is kind of Kant as well, I guess. And this leads to the early Leibniz error, or the same error that the good old fashioned AI practitioners had, which is that intelligence could be carried out by just having a series of programs of strictly logical deductions or inductions. That doesn't work. This is why good old fashioned AI never panned out. And the reason is that you can't start out like in math, with propositions that are self sufficient. Even math is not self sufficient. But let's pretend for a moment and just move from there and kind of do an algebra in order to work various things out. When your propositions are not airtight and when you're looking only at regularities and patterns, this good old fashioned AI idea simply cannot work. And that's why we never got it to work. Let's move now to some of the artificial life experiments that that I began playing with at the end of 2023 and my team and I published in June of 2024. So just about a year ago, I think some of you, many of you perhaps have heard of these. They're in the what is Lifebooks? And I've talked about them a few times. The basic setup here is to try and get self replication to get, you know, abiogenesis, the emergence of life from non life, to happen in a purely artificial life system. Okay, so the setup is to begin with a minimal Turing complete language. I used Brainfuck because I really liked the idea of being able to talk at a conference and say brain fuck over and over. And I'm fundamentally 12 years old on the inside. But also because it very closely models the Turing machine. It's a minimal programming language. There's only eight instructions that looks very Turing machine like and moves the head back and forth. I should say that in its original version, Brainfuck is not embodied computation. It has basically a separate data tape and code tape. And that means that it cannot make a copy of itself. So I made a couple of modifications to Brainfuck that actually reduce it from eight instructions to seven in order to make it embodied, meaning that as it works on the tape, it is able to Read its own code and write. And write its own code on that tape as well. There's no separate console. There's no separation between the data tape and the instruction tape. For those of you who are unfamiliar with brainfuck, there is hello World in it. I'm sure you've already figured out how it works by just looking at the program. I actually still haven't, I have to admit. By the way, this is actually the French brainfuck page because I thought it was better, but translated into English. It's funnier to read it that way. These are the eight instructions. You know, the first four are move the head one step to the left, one step to the right, increment the bite at the head, decrement the bite at the head. We're already halfway through. There is an input and output instruction, which in this case really just copy from one head to another. And there are jump instructions, open bracket and close bracket in order to be able to make loops. And that's it. That's all Brain Fuck is. So how does the ALife experiment work? The ALife experiment is called BFF. The first BF stands for brain Fuck, and the second F. You can draw your own conclusions, but you start off with a soup of. I actually generally use just 1,000 1024 tapes. That's enough for this experiment. So the tapes are of fixed length. They're of length 64, and they begin random. So just random bytes. Now, if a tape is random bytes, that means that only 1 in 32 of them or so are even valid instructions. Most of them are no ops. A no ops will just be skipped over like in most programming languages. So this is what those tapes look like in the beginning. And you can see that I'm not printing the nops right. So that's all the blank space. The operations are quite sparse. On any given tape, you only have an average of two instructions or so. And then the procedure is to pluck two of these tapes out of the soup at random, concatenate them end to end. So you have 128 bytes and then run, and then after running, pull them back apart and put them back in the soup and repeat. That's it. So it's just that over and over, that's the entire experiment. So I'll show you what happens on my laptop. After a few million interactions, magic happens, which is that you go from noise to programs. You start to see complex programs appear on these tapes. And this is quite wonderful because these programs take real effort to reverse engineer when you study them. It's like studying that hello world program. They're functional in the sense that they really do something. And it's not trivial to figure out how they work in order to do that. Okay, what are they doing? Well, they're definitely copying themselves or each other somehow. We know that because this is a histogram. And you could see in this case there were 8,000 tapes. There are 5,000 of the top one, 297 of the next one, and so on. So there's clearly copying going on. And there's this ecology of programs all copying each other, which is just wonderful to see. That's that emergence of life in this very functional, minimal sense from randomness. A part of this is very easy to understand. Why do these things emerge? Well, because something that copies itself will be around forever, and something that doesn't copy itself will be copied over by something that can copy itself. So inherently, something that can copy itself is more stable than something that cannot copy itself. So it's really just the second law of thermodynamics, but doing something unexpected, which is creating something more complex because it's more stable, rather than something less complex, which is less stable. This idea that stability doesn't necessarily mean low complexity was worked out in some detail by Addy Pross, the organic chemist in another book called what Is Life? He calls it dynamic kinetic stability. Meaning usually we think of stability only in terms of fixed points in a phase space. But a cycle can be even more stable than a fixed point. Of course, for these cycles to work, you need an input of free energy for reasons that we've already gone into mystery, mostly solved, but actually mystery not fully solved, for reasons that I will show in a second. But just to give you a sense of what this transition looks like from non life to life, it's very dramatic. In the beginning, these interactions only involve there are only a few instructions in the soup. It's a Turing gas, as Walter Fontana would have called it. When you do the join and you run only two operations run in any given interaction, on average, as you'd expect. And that's what it looks like by the end in this particular run. And 1374 operations on average are running per interaction. So the soup has become intensely computational. There's been a transition here, and there's a lot more code than 1 in 32 bytes, as you can see. This is what that looks like visually. This is the most exciting plot that I've made in the last few years, and it's the one that's on the COVID of the book. So what I've drawn here are 10 million dots. It's a scatter plot of interactions. The X axis is time, and the Y axis for every dot is how many computations took place, how many operations took place in that interaction. And you can see that in the beginning, it's not very computational. And then a sudden transition takes place here at 6 million interactions, and it becomes intensely computational. It looks like a phase transition. In fact, it is a phase transition. You can also see that in the entropy of the soup. So here I'm just estimating the entropy of the soup by zipping it and looking at the size of the zip relative to the whole thing. You can use any compression algorithm you like. In the beginning, it's uncompressible, so it's a gas in that Turing gas sense, because all the bytes are random. And you can see that there's a dramatic change. And suddenly it becomes extremely compressible right at that transition moment. And of course, this becomes compressible because everything is copying itself and each other. So if things are copying themselves, then we know that they'll become very compressible. But it's cool, because if we think about what the phase of matter is on the left, it is just like a gas. Nothing is correlated. What would we call the phase of matter on the right? It's not liquid, it's not a solid. It has structure, and it has structure at every scale. I think you have to call that phase of matter life. It's a functional phase of matter. It means that its parts are different from its other parts. And if you zoom in or out, you see more structure. So it's what David Wolpert would call self dissimilar. It's not a fractal, it's more like a multifractal. I'll explain why in a moment. Okay. How long does it take this transition to happen? Well, the answer is it looks more or less like an Erlang distribution, or a little bit more precisely, like this distribution I call a lock pick distribution, which imagines that there are steps that have to be undertaken, and those steps have a long tail distribution of difficulty. And how many steps does it take? Well, the answer is 12. It takes 12 steps, just like getting sober. I suppose. This is a fit of the empirical to the erlung and the lockpick distribution. It's a little hard to see, but the lockpick is a bit better than erlung. Erlung assumes poisson, lockpick assumes long tailed, but it's a Process phase distribution. And what this tells you is that there are stepping stones here. You can't get that transition to life immediately. So something interesting must be going on here on the left. Other than just randomness, it takes multiple things happening in order to get to that point. In this case, it happens somewhere between 1 million and, let's say, 7 million interactions. Okay, so this all suggests that pretty much any universe, by the way, that has a source of randomness and can support computation will evolve life for this simple dynamical stability reason. But the big mystery is why does it appear to get more complex over time? You might have seen my little video that we saw some programs emerge and then we saw them densify more instructions appeared. Even more fundamentally, why does this work even without mutation? I didn't mention, but in the original version of BFF I added some random mutation because we're all taught in school that the way evolution works is chance and necessity. You mutate things, you're throwing spaghetti at the wall and whatever sticks is what does better. And so you need a source of spaghetti. But if you do this entire experiment with the mutation rate cranked all the way down to zero, you still get the same exact phenomenon. And that is very mysterious because if you crank mutation down to zero, you should have no source of novelty, you should have no evolution. Why do you still get this apparent complexification even with zero mutation? So let's go into some of the theory of this. By the end we have a replicating entity. It can engage in standard sort of population evolution dynamics. This is the kind of differential equation that one generally writes for this sort of thing. It's a very general ansatz. This is for species. I let's say there are n species. They could be chemical species, they could be biological species, whatever. Here's a classic example of such an ansatz. This is the Lotka Volterra equations for predator and prey, which I'm sure many of you are very familiar with. They were co invented or invented independently by Alfred Lotka and Vito Volterra near the beginning of the 20th century. This is what the classic Lotka Volterra equations look like. There are two species. There's a prey species and a predator species. And those four terms are reproduction, getting eaten, eating to reproduce, and background death rate. So if you've got those four terms, you get these nice oscillatory solutions between your predators and your prey that arise. Okay, so this is a slightly more general form of those Lotka Volterra equations. There is a linear part which we'll call rx. And in Lotka Volterra that linear part is diagonal, so the wolf can't turn into a rabbit, the rabbit can't turn into a wolf, so the reproduction is diagonal. And then there's also a bilinear term, which is the part where predation, competition and the fact that niches are finite gets implemented. So the right part is suppressive, the left part makes things grow, the right part makes things squish, squish down, keeps them finite. But this can't be the whole story of evolution. Why can't it be the whole story of evolution? Well, of course, because it's closed, ended. We only have two species here. It doesn't matter how long you run this damn thing, you're not going to get a third species and you're not going to change the design space either. You can have very complicated terms in here that allow finch beaks to adapt to different environments, but you have to have the space of finch beaks predefined before this equation can even be made to work. So this doesn't answer the question of how evolution gets started. It doesn't answer the question of what happens afterward other than optimization to niches. So now we bring in another Eastern European, Dmitry Sergeyevich Mereshkovsky. So he's the one who first came up with the idea that maybe mitochondria engaged in some kind of simulgenetic event in order to end up inside other single celled organisms to make eukaryotes. This was popularized and proven to actually be the case by lynn Margulis in 1968, one of the really great papers in biology from the 20th century. I'm sure many of you are familiar with. This is that paper, sorry, 1960, on the origin of mitosing cells. So she's the one who proved that eukaryotes were actually a fusion between two different kinds of prokaryotes and popularized this term that mere invented symbiogenesis. Okay, so could symbiogenesis be happening as a source of novelty in bff? Yes, that is the source of novelty in bff, and indeed that is the source of novelty in evolution, period. This is something that Lynn Margulis believed, but that had not been widely accepted by the biology community even by of her death in 2011. So she had a much more expansive idea about why symbiogenesis was important. Only the particulars of chloroplasts and mitochondria had been accepted. So the way we can look for symbiogenesis in BFF is to look for replicators emerging before that phase transition. And if you look for them, if you just look for stretches of bites that are getting copied during those interactions, you find such stretches of bytes. They begin short and kind of crappy, unreliable, but they're there from the beginning. Every time you have a single copy instruction, after all, one byte is getting copied from somewhere to somewhere. So almost by definition you have at least 1 byte long sequences that are getting copied right from the beginning. So let's just call them replicators. There are replicators there from the beginning. Now, if you have these one byte replicators that are copying themselves back and forth now and then once in a while they will come into conjunction and two of them will copy better as a group than the two of them copied on their own. And when that happens, then they'll start to copy as a group. And that is a symbiogenetic event. Basically, the reason that even without mutation you get these complex programs arising is because of these fusion events between smaller replicators. Can one build symbiogenesis into an equation like this one? For Lotka Volterra, you can. This is our statistical physicist who came up with the right kind of term to write mathematically for describing how symbiogenesis works. He wrote down an equation for the coagulation of polymers. So this is Smolkowski coagulation. This is what happens when clouds form. It's what happens when gelatin sets in the fridge. So the idea is that you have, let's say polymers that begin as monomers, one monomer, another monomer, they stick together. And now you have a dimer, and now the dimer and maybe another monomer stick together and you have a trimer. Two trimers stick together, and now you have a hexamer, and so on. These are the equations for that. This is the mass balance equation. It's very simple. There's a merger gain term and a merger loss term. The merger gain term, which scales like the densities of the two things that are coming together and the product of those with some merger kernel K is increasing the population of cluster K, which is of length I plus j. And then you have to do the balance of that. Every time you have two things coming together to make a new one, you have to then subtract their populations I and J. And that's what the right hand side is about. It's the loss of things that have merged. So you put those two things together and you get a stochastic differential equation for mergers in a solution. And by the way, there is a phase transition associated with small husky coagulation. It's called gelation. And it's exactly what happens when you put Jello in the fridge. And it sets, basically, if things are sticking together, and if they stick together with a scaling exponent that is greater than one, then you get this finite time singularity in which the things that stick together diverge to infinite size and the whole thing sets, no matter how big it is. And that's how Jello sets. Could that be gelation? Yes. The short answer is that is gelation. That phase transition that we see of the emergence of life is a gelation phase transition, according to a generalization of Smolhovki coagulation, to this case of bff. Strings coming together. If you think about inanimate and viral replicators as being replicators that are not self contained, in other words, where the code that runs is not fully within the code that is actually getting copied, then you notice something interesting. So what I'm calling here, an inanimate replicator, and very much in scare quotes, is code that copies something fully outside itself. In other words, the code that runs in order to do the copying is disjoint from the thing that gets copied. Are there such replicators in the real world? Of course. That's what water is. Water is a replicator of some kind. It gets made by stuff, but the stuff that it gets made from, like water, is not a part of the running process. It is a part of the running process that makes more water in some cases, but it's. It's not part of the code. Let's say viral is the case in which the code and the thing that is copied overlap. So in other words, some of the code that does the copying is actually some of the stuff that gets copied, but the code is not fully contained by what gets copied. So this is an incomplete replicator that would need to cooperate with another replicator in order to reproduce. So that's what I mean by viral. In the beginning of bff, all of the replicators are inanimate and viral. The great majority are inanimate and a few of them are viral. A few of them happen to copy one of those bytes that is actually an instruction that is doing the copying. But as you move toward the time of gelation, which I've normalized to one here, you can see that cellular replicators suddenly emerge. So they can't emerge before about halfway through the run, and they shoot upward at the end. And that's really interesting because that tells you that the Moment of a cellular replicator where the machinery for copying yourself is part of the thing that is copied, emerges through the symbiosis or the symbiogenesis of inanimate and viral replicators. Okay, so a full equation would have two terms. It would have this reproduction and Laka Volterra type term, and it would have a merger or Smolhovsky type term. One on the left is normal population dynamics. That's normal Darwinism. And the one on the right, you could think about the left as evolution and the right as revolution. That's the moments when things come together. Now, the population dynamics part for BFF looks like this. It's a little bit more complicated, but it has the same basic form as Lotka Volterra. There's a linear part on the left. I'm just writing that as a matrix R I, J operating on the whole thing. And on the right, the reason that looks a little bit different from Lotka Volterra is that when something gets copied, it overwrites other stuff. So now we have to say, well, how does that suppress the populations of everything else in the soup? In order to figure that out, you have to look at niches. What are the bytes where something gets copied? And the overlap between the niches of two replicators tells you how much one thing getting copied, how likely it is that that will overwrite something else that shares its niche. Okay, the symbiogenesis part is a bit of a mess, so I'm not going to go through it. I hope that's okay. But it looks just like Smolkowski, just gnarlier. The reason that it's gnarlier is because Smolkowski has only binary fusion between two parts. And in bff, sometimes a bunch of things come together. So you have to take into account these kernels that have more than two parameters in them. Also, when things come together, they don't necessarily look like the sum of the things that came together. You could have something that is 3 bytes long or something 5 bytes long come together, and the result that copies itself is only 2 bytes, 1 byte from each one, or anything right along those lines. So to account for those complexities, you end up with a much more complicated K term, but it's essentially the same as Smolhovsky coagulation. To prove that this kind of symbiogenesis is needed in order to get these complex programs, you can do a very simple intervention, which is when you're interacting two tapes, you can sort of do it in a sandbox before committing. And in the Sandbox, you see whether a new replicator arises, and if so, what replicators is it made out of. In other words, when you look at the source, you can see whether any of those source bytes were actually the outputs of copies of some previous replicator. And if so, then you have a tree. You have an ancestry tree for that replicator. That means that you can think about the depth of such a tree, how many things have come together, and you could limit the depth of that tree. You can say if the tree depth exceeds 10 for a new replicator, then I'm going to actually not do this interaction. I'm going to take them back apart, pretend it never happened, put it back in the soup and try again. If you limit the depth of the tree to say, 24, then the number of operations that you have to block, the number of interactions you have to block is actually very small. You only have to block one in a thousand operations. But that one in a thousand operations is really important. As it turns out, if you block those, no gelation will happen. You need at least tree depths of 20 or so in order to get these complex programs. So this is very nice proof that symbiogenesis is what is needed in order to get to these complex tapes. When you do that blocking, you end up with sort of logistic curves for the populations of all the replicators in that soup. They go up and then they saturate and stabilize. That's fun because it lets you do a little bit of math. So as you can see, not only do things go up and saturate, but then there's some random oscillations. And those oscillations can be correlated. Sometimes you can see the two of those populations go up and down together. So that means that they're maybe collaborating with each other. And sometimes they go in opposite directions. They're anti correlated and that means that they're competing with each other because one is overriding the other, for instance. So that's what one would expect from off diagonal production and competition from those equations I wrote earlier. And if you linearize the dynamics around that steady state, then you can sample the correlations in those population fluctuations and you can reconstruct the matrix R. I'll skip the details of how one does this, but this is a classic fluctuation analysis. You solve the Lyapunov equation and you get a Jacobian. And from that you get the matrix R. And the matrices R look really cool. First of all, they have a strong diagonal that tells you that by and large things replicate themselves Just as you would expect from Lotka Volterra. But there's some other stuff going on here as well. Aside from that dominant diagonal of self replication, there is some negative stuff off the diagonal and some positive stuff off the diagonal. The negative stuff off the diagonal you can see looks largely symmetric about the diagonal. And that's as you would expect too. Basically, if A competes with B, then B competes with A. Two things that are fighting for the same niche are in a kind of zero sum relationship with each other. But the cooperation part, where something helps something else, is not symmetric. And that's as you would expect too. Just because A helps B or enables B doesn't mean that B enables A, or at least not directly. So there are complex cycles in this graph on the right of co dependency or enablement. So a negative component is symmetric, positive component is asymmetric, and there's this big diagonal. Do the submatrices that are about to undergo symbiogenesis have any special properties? They do. So in other words, if it's these, let's say, four rows and columns that are about to undergo symbiogenesis, you can ask what are the eigenvalues of that matrix, of that submatrix? And it turns out that they are generally cooperative. So essentially, if you were to pick random rows and columns from this matrix, then you get high dimensional picture of the rank of the matrix. But when you look at the ones that actually combine, it's much lower rank, they're already working together. So in other words, there's a relationship between the R and K parts of this equation. Symbiogenesis happens among guys who are already working together. Not all the same, not independent, cooperative. Here's another really interesting thing. If you look not at the R matrix, but at the Jacobian itself, then you can find the signs of imminent instability in it, of when it's about to pop, when it's about to go run away and gelate, or gel. You don't say gelate, you say gel. Right? So in particular, if you block the depth of the possible trees to a low number, then the eigenvalues of the Jacobian are always negative, meaning that the system is stable. But as you look at larger depth ceilings, you find that more and more of these leading eigenvectors, or the real parts of those leading eigenvectors pop positive. And that means that the system is about to blow. You can keep it from blowing for a while by keeping that merger clamp on, but it tells you that essentially the more you evolve these things the more they begin to cooperate with each other and the more incipient symbiogenesis is about to happen. And that's what leads to this phase transition. All right, I just want to put a little plug in for what I think could be a really beautiful missing link between the kind of algorithmic information theory that Hector Zanil was talking about and the assembly theory that he has somewhat slammed with a couple of papers that he has written. But as those of you who have followed that might know or might realize, from what I've just talked about, there's a very close relationship between what I've just been describing and assembly theory. It's things coming together to make bigger things. But the assembly theory proponents have not really talked about the computational nature of what they're doing. And in this, I fully agree with where Hector is coming from. And the way that those connect, I think, is by starting to look at things like conditional Kolmogorov complexity of the things that are coming together. So I think this is a construction point for us to maybe reconcile those two different pictures. All right, so symbiogenesis is what gives you complexification. That, in turn, is what gives evolution its arrow of time. In classical evolution and Darwinian evolution, there's no reason that things should become more complex over time. They might simplify, they might get more complex. It doesn't matter. But with symbiogenesis, we know that things get more complex because if A can replicate itself and survive into the future, and B can replicate itself and survive into the future, when they come together, you suddenly need A to replicate itself and B to replicate itself. And there's some additional information that has been added, which is how the two fit together. And those extra bits of information that keep getting added to the program of what is the large replicator, they don't come from mutation. They come from the fact that things encounter each other randomly in order to possibly undergo that symbiogenetic event. So it's actually the thermal randomness of the fact that we pluck two of these guys out of the soup at random. That's the information source, if you like, or the noise source that is selectively turned into algorithmic information by the symbiogenetic process. Jor Safmeti and John Maynard Smith have written extensively about these major evolutionary transitions in which symbiogenesis results in large, novel forms of life like eukaryotes, multicellularity, and so on. And I think this work is great. But the flaw is that they're only talking about eight events or 12 events. And if what I'm saying is true, then this is just the tip of a gigantic iceberg. Basically, it's symbiogenesis all the way down. Most of these symbiogenic events are much more uneven. They're maybe just a little bit of something getting incorporated into something much bigger. But that is the source of novelty in all of evolution. These are just the most dramatic cases that involve really big, visible stuff happening. So is there evidence for these smaller symbiogenetic events in biology? Lots. There's lots of evidence for it. So I don't have time to go into it in any detail, but if you look at just the human genome, you find that only 1.5% of it codes for our proteins. And lots of the rest of it is transposons and other endogenous retroviral elements of various kinds that involve viruses, whose ecology is our own genomes and that reproduce inside our genomes and sometimes jump species, resulting in weird shit, like a quarter of the cow genome being a retrotransposon that also lives in lizards and salamanders and stuff. And so when you start to look at that, you realize that genomes are fractal, and it's replicators made of replicators made of replicators, just as I've described. Not these kind of fixed design, space, and evolution only happening in its usual way. It's not just horizontal gene transfer in bacteria. This symbiogenetic picture, I think, is the engine that produces novelty throughout all of life, including big, complex animals like us. There's more and more evidence in the last decade of things like this going on. For instance, the ARC virus was endogenized in the mammal lineage, and you can find it in our brains. And it turns out that if you knock out the ARC virus in mice, they stop being able to form new memories. Clearly, the ARC virus is doing something important for us, and that's a source of novelty. That was an endogenized virus. Similar, the mammalian placenta is formed by an endogenized virus that fuses cell membranes together and so on. There's a definition of life that comes out of this life, and I said this in the panel yesterday, is an embodied autopoietic computation arising and complexifying through symbiogenesis. It's not just neuroscience. That's computational. Life was computational from the beginning. It gets more computationally complex over time through symbiogenesis at many scales. Because, remember, if life is a computer from the start, then every time things fuse together, you're making a more and more Parallel computer. Those computers have to be not only running the code that model themselves and reproduce themselves, but that also do something about modeling the other and figuring out how they interact or work with the other. And this means that an ecology of functions is building up through massively parallel computation that becomes, if you like, more and more intelligent with every one of these fusions. And since symbiogenesis makes the computation massively parallel, that implies that intelligence and life are very, very closely connected. Which is why I ended up with the book what is Life? As part of the book what is Intelligence? When you're not only using that intelligence to model yourself, but also to model your environment, which by the way, includes others, most importantly, then that's intelligence. And that means that life was intelligent from the start. And the moment that modeling of others begins, what we call in larger, more complex animals, theory of mind becomes fundamental to the way intelligence develops. So these are really simple simulations that show how just persistence allows the modeling of an environment to turn into learning chemotaxis in these fake bacteria. But of course, in real life, you're not only learning about an environment that exists in isolation, like the sugar crystal, but actually all of your friends, right? The moment you're reproducing, the greater part of your environment is actually all of the other things that even your own reproduction is creating. Life is never single player. Things like intelligence explosions in our lineage, in the hominins and in cetaceans and in bats and in a variety of other species, are exactly this kind of runaway modeling of others, resulting in growth of brains and growth of groups. And therefore that when we think about the growth of advanced intelligence in human societies or human brains, it's really that same sort of symphogenetic process happening at a much higher level. Let's end there and switch to questions. The world moves fast. Your workday even faster. Pitching products, drafting reports, analyzing data. Microsoft 365 copilot is your AI assistant. Assistant for work built into Word, Excel, PowerPoint and other Microsoft 365 apps you use, helping you quickly write, analyze, create and summarize so you can cut through clutter and clear a path to your best work. Learn more@Microsoft.com N365 copilot. Eczema is unpredictable, but you can flare less with Epglis, a once monthly treatment for moderate to severe eczema. After an initial four month or longer dosing phase, about 4 in 10 people taking EPGLIS achieved itch relief and clear or almost clear skin at 16 weeks and most of those people maintain skin that's still more clear at one year with monthly dosing. Empglis Lebricizumab LBKZ, a 250mg 2ml injection, is a prescription medicine used to treat adults and children 12 years of age and older who weigh at least 88 pounds or 40 kilograms with moderate to severe eczema, also called atopic dermatitis, that is not well controlled with prescription therapies used on the skin or topicals or who cannot use topical therapies. EBGLISS can be used with or without topical corticosteroids. Don't use if you're allergic to ebglis. Allergic reactions can occur that can be severe. Eye problems can occur. Tell your doctor if you have new or worsening eye problems. You should not receive a live vaccine when treated with ebglis. Before starting ebglis, tell your doctor if you have a parasitic infection. Ask your doctor about eglis and visit eglis.lilly.com or call 1-800-lilyrx or 1-800-545-5979. I think there are multiple different ways to represent the symbiosis. I think in the real biology, we maintain those high record structures and that there are fundamental mathematical differences. How you treat those symbiosis. Do you have any insight how we can implement that? Yes. In biology, we often sort of reify one particular level of detail and we say these are the life forms. And maybe there's a symbiosis between, let's say, you know, algae and a sea slug. But we still think of the algae and the sea slug as separate, and we think about population dynamics within that rather than modeling them separately. Is there a reason to prefer one scale or another? For me, one of the lessons, the reason that I spent so much time on the relationship between R and K, is that you can always move something from R to K and back. Lynn Margulis famously said we're just colonies of bacteria, some of which live inside each other. And that's true. You could describe us as just colonies of bacteria. But the reason that it's useful to move up a level of detail is because humans also reproduce as a unit, hence the mets. And there are a lot of things that you can learn about when you study at that higher level. A lot of abstractions you can make from a computational perspective that are hard if you're only modeling at the lower level. So I don't think that there's any one layer level that is true. And we have lots of boundary cases like lichen or like colony insects, where you can model the entire colony or you can model the individuals. I don't think there's a right answer to those two. So do you keep a block of rows and columns in R that always end up or mostly end up getting copied together, or do you add a new row? It's actually a coarse graining choice and you can make either one. Symbiosis is not symbiogenesis. What is the thing that you would claim is kind of like a good insight for how A and B stop being A and B in symbiosis and they become something else? The fact that phase transitions don't come from R alone, but you have to look at K. Now you do get these runaway modes which tell you that something is about to happen. But in order to understand that phase transition, in other words, in order to see that a major evolutionary transition has occurred, to put it in biological terms, you actually have to understand the physics of K. It's only by understanding the physics of K that you can do the theory that lets you predict and understand what is going on right here. So if you model a body as just a bunch of bacteria, then it's not wrong, but then it's invisible to you that something amazing happened when we became multicellular or when eukaryotes formed out of bacteria. So this allows higher order modeling. And in particular, by the way that higher order modeling, if we just take the subjective perspective for a moment, if you are one of those bodies, if you are one of those people, then you're not going to survive very well. If you're only modeling other people as collections of bacteria, you have to build higher order models of them, because that becomes an essential part of your umwelt and you have to simplify or coarse grain the world in order to build a model that is ecologically relevant to you. So I'm kind of mixing here of subjective and an objective perspective. But the subjective perspective ultimately is super important. It's a little bit similar to like why do we need temperature and pressure in physics? Those don't exist if we just look at the microscopics. But it's only by coarse graining and looking at the larger scale that you can understand thermodynamics. And in the same way, it's only by zooming out and looking at the symbiogenesis that you can understand the dynamics of transitions, of phase transitions, METs and smaller and coarser grained models, where higher orders of things emerge. Well Actually, that's exactly what I was doing like 30 years ago. Right, exactly. That's why I began with your nearly 30 years ago paper. There was a big problem, right? It's always symbiosis is possible, like we burn wood. Then it becomes complicated coupling with each other. However, the function itself is not becoming complex or because higher order things function itself is. Just stay there. I have three answers, I suppose, depending on which way we talk about it. So first answer actually comes from software engineering, so composition, or from mathematics for that matter. Composition of functions is symbiogenesis as I've described it. And when you compose two functions to make a higher order function, you are making something more complex than the primitives. And what I hinted at with Eric EL Moll's beginnings of conditional Kolmogorov complexity can quantify that sort of compositional complexity. You could find signatures of it in. If you don't want to look in DNA, you could look in GitHub at the way every time somebody writes some code it begins by importing a bunch of other things and combining them. And you do see a tendency toward complexity now that is constrained by energy. The more complex a thing you make, the more free energy it has to use. But now you get some of Chris Kempis beautiful work in which you see that there are energetic benefits to teamwork. So the way the scaling laws work for this, which are also environmentally dependent. I don't know, Chris, if you got to the snowball Earth type stuff, but there were certain very specific conditions at certain points in the Earth's history that became favorable for eukaryogenesis, if I'm remembering correctly. And so there are some external conditions as well. But the other cool thing is that when you start to have more complexity in the computers, when you start to have massively parallel computation, that greater intelligence also unlocks new energy sources and that gives you a bigger budget to play with, which in turn allows the next MET to take place. So I think there's an energetic perspective, there's a compositional perspective, a combloric perspective, there's a scaling law perspective. That can all come to rescue that question. But we should talk more about it. I'd love to get into this in more detail. It's essential for life to have some functionality. When I think formally of the brainpack programs, they're strings in which gain their functionality by external goal, AKA the cpu. I agree. So I made claims that maybe sound contradictory, one of which was that von Neumann is embodied computation and different from Turing in that sense. But on the Other hand, that BFF looks very much like a Turing machine. And yet I also said it was embodied because I made one tape. So even the question of whether something is embodied or not is a little bit perspective dependent as well. Because in a von Neumann system, for instance, there's of course the same rule operating at every pixel. You can ask yourself the question, is a computer the thing that I make with lots of parts of the kinds that von Neumann designed, or is it just the operation of a single pixel? The usual answer is to say what happens at a single pixel is just the physics of that world. But what constitutes the physics and what constitutes the computation is actually a movable boundary. So embodiment is essential in the very minimal sense that you need to be able to operate on the thing that is going to. You need to be able to make quines, essentially, right? And in ordinary brain, fuck, you can't make a quine because the data tape is separate from the program tape. But you bring them together, you're now in the same realm as a cellular automaton, albeit with a different core screening of what you consider to be the physics and what you consider to be the code. Now, I mean, in our world, we know that it's possible to build computers, or else we wouldn't be able to build computers and we wouldn't be here either. But what constitutes the physics, if you like, the physics that makes up computers itself had to evolve. We began as, I guess, nothing but quantum field theory. And then things came together into particles and the particles come together into atoms, the atoms come together into molecules, and so on. Those are essentially what I would call the inanimate replicators in this system. And there is an important phase transition when those suddenly form a rich enough set that not only do you have an autocatalytic system, as Mulder Fontana would have said, but also that you can form a Turing complete instruction set and therefore open the door to generality of computing. I hope that makes some sense. All right, I think we're. Yeah, I'm afraid we have to wrap up, so let's thank Poise once again. Thank you so much for this amazing. This amazing talk. Great question too. New Year, new Me. Cute. But how about New Year, new money? With Experian, you can actually take control of your finances, check your FICO score, find ways to saved and get matched with credit card offers, giving you time to power through those New Year's goals. You know you're gonna crush start the year off right. Download the Experian app based on FICO scoring model offers an approval not guaranteed. Eligibility requirements and terms apply subject to credit check, which may impact your credit scores. Offers not available in all states. See experian.com for details. Experian this Precedence Day Upgrade the look of your home without breaking your budget. Save up to 50% sitewide on new window treatments at Blind Spot. Blinds.com makes it easy with free virtual consultations on your schedule and samples delivered to your door fast and free. With over 25 million windows covered and a 100% satisfaction guarantee, you can count on blinds.com to deliver results you'll love. Shop up to 50% off site wide, plus a free professional measure during the President's Day Mega Sale happening right now@blinds.com terms apply.
0:00