Darwinian Objectives
Evolving AI, Survival, and The Meaning of Life
Evolving Intelligence
I thought it was worth documenting an epiphany I had recently about how intelligence evolved. The long-term goal of AI research is to create human-level (or greater) intelligence, and we know this is possible because it has already happened once before, via evolution. While the brain is too complex for us to currently understand in full, one possible approach to replicating its ability is to look to the underlying process which created it. Perhaps we could construct some simulation of life or nature and let it run for a very long time, and some simulated blob of cells would appear at the other end, resembling intelligent life.
This is not an especially original idea (the entire subdomain of genetic algorithms is based on this principle), but it is a very interesting one to try thinking through. Hypothetically, if we were to construct some evolutionary process to evolve intelligence- what would it look like? If we follow the format of a typical GA, we need to define a fitness function. What exactly would we be trying to optimize for?
If we follow the typical understanding of "survival of the fittest", we don't actually need to put in much effort here if we don't want to. We can assume that, on average, the simulated beings in our hypothetical "genetic algorithm of life" will be more prone to survival if they are somehow more capable than their peers. One axis of capability is intelligence, so if we simply design a preference for individuals that survive, we are indirectly reinforcing intelligence. We just need to simulate a few billion generations of life and death, and whatever we end up with is probably more intelligent that what we started with because, well, it made it that far.
So far so good. We know that from generation to generation, we are just going to prefer "the members that survived." But what does that actually look like? Within a single generation, what do our simulated beings actually do?
Survival of the Fittest
At this level, within a single generation (or period of time or however we have set up our god simulation), things get a little weird. In a typical GA, each population is a set of solutions and we prefer the fittest ones- the ones that best solve the problem. In this case the fitness function is actually just survival, so we just pick the ones that... don't die?
This seems a little reductive, because if things are working correctly then our population is itself made up of partially intelligent beings, and they are capable of pursuing goals. That is, the beings themselves are doing things with their life, and actively pursuing survival within their lifespan. We end up in a sort of meta-algorithm, preferring the members of the population that were able to best achieve their own survival, or live their best life, or self-actualize, or something.
Note that this doesn't require especially high levels of intelligence to be a concern. Even the most primitive lifeforms, operating entirely on instinct, might pursue nutrients and avoid danger. This seems a critical aspect of life itself: life "wants" to survive. The members of the population are actively trying to do something within their lifespan, and more than raw potential this is what evolution seems to really be rewarding: the ability to execute the task of "stay alive".
So somehow, we need to encode within our simulation the goal not only of the overall algorithm, but the goal of a single lifespan. What exactly should a semi-intelligent being be trying to do over their lifespan? What is the purpose of their life?
The Meaning of Life
The most obvious solution to this problem- that the individuals are trying to stay alive- seems wrong when you start to think about it. If individuals simply pursued their own survival, life wouldn't last very long. All the primitive algae in the primordial swamp would just float around, eat stuff, not die for a while, and then they would succumb to old age and, in fact, die. Life as we know it has ended.
This is a conundrum. Death is inevitable, so individuals in nature are not purely trying to survive for their own sake, which is futile in the long run. Instead, they are trying to survive for the benefit of their species. They are trying to reproduce, create families, form societies, and ensure that their children have a better life in the primordial swamp than they did- which is absolutely bizarre when you think about it. Compared to the dog-eat-dog brutality of nature we are trying to mimic, we seem to have this encoded altruism in each individual. While they will not live past some point and therefore could not benefit in the long term, individuals try to make sure that they create offspring, that those offspring survive, and that those offspring prosper. Even more than survival of the self, this seems to be the motivator in nature- survival of the species.
From an algorithmic point of view, this is extremely odd. If we imagine our two-tiered algorithm, at a high level (generation to generation) we prefer the most capable individuals (survival of the fittest). At a lower level (within a generation), the individuals are striving to keep themselves alive but also ensure survival of future generations of their species, to no benefit that will be realized within their own lifetime. It's almost as though the goal of the individuals in this algorithm is to make sure that the higher level algorithm has something to do. This genetic algorithm's members strive to ensure the perpetuation of the algorithm itself.
Existential Crisis
At this point I hit a mental brick wall for a long time, and possibly an existential one. The more I thought about this the weirder it seemed. Why would individuals in nature optimize anything at all, given that in the grand scheme of things their existence is almost instantaneous? What would an individual's objective function even look like? Unless we manually intervene and encode a desire to reproduce, what more general objective would actually result in this behavior? Furthermore, what does this mean for our overall objective- is our fitness function really just survival of the individual, or something more complex? Do we need a "species fitness" as well?
Maybe we, as products of the algorithm of evolution, are not equipped to reason about it. Just as we do not fully understand the human brain, maybe we cannot understand life? Not to mention, we assume that human intelligence is the current best product of evolution. Is that true? Does natural evolution even have "an objective" in the first place? Why am I thinking so much about this?
Survivorship Bias
About two days ago I figured it out, totally at random. I was thinking about this all wrong.
Suppose that this weird meta-algorithm entanglement wasn't happening, that individuals in nature were not trying to perpetuate their own species. What would happen? Nothing would happen. Life would just stop after a few generations, since the "next generation" would be of no concern. The individuals die off, the higher-level algorithm has nothing to do, and evolution grinds to a halt.
And yet, here I am today writing a blog post, so we know this didn't happen. In fact, the only way that evolution could have kept churning for billions of years is if the long-term survival of the species was prioritized by individuals. Evolution exists, and therefore species-level survival must be prioritized. Since it is prioritized, evolution exists.
To put it another way, there is nothing special or magical about this property that individuals perpetuate their species, no venerable altruism baked into DNA that looks beyond our own lifespan. In reality, there have probably been millions of species that did not have this instinct, and they are simply dead now. This is just a basic requirement for evolution to function, so species that do not have this particular drive are lost to time.
To self-diagnose, I suffered from a quite literal case of survivorship bias. I was only considering the lifeforms we know about, the ones that actually stuck around long enough to be influenced by the process of evolution. Most mutations are harmful, and most potential species just never made it that far, so it is easy to forget about them.
If Intelligence Evolves in a Forest...
... and no one is there to recognize it, did it actually evolve? In a strange way, evolution presents us with a miniature Anthropic principle. Evolution on earth seems to have special properties simply because, well, here we are. Evolution could have happened many times, on earth or elsewhere, but it doesn't really count if it doesn't get far enough that its own products recognize what is going on.
To return to the original question- how to design a genetic algorithm that creates intelligence- my answer for now is don't overthink it, just make sure you run it over and over and over. If the algorithm converges too fast, and evolution stops, then that is probably par for the course. But at some point, by shear luck, a run will go for a while. It will self-perpetuate. The simulated beings will develop some strange mutation that encourages the survival of the algorithm itself. And then... you just see where it goes!
Recent Posts:
Flowing with Fewer Steps
Shortcut Models Notes and Review
December 12, 2024
Going with the Flow
Notes on Flow Matching (Policies)
December 9, 2024
Modeling the World
RSSM & TSSM Notes and Experiments
December 1, 2024
More Posts