"A predator must predict actions that will get the prey into its stomach; the prey must predict the predator’s behaviour to stop that from happening. Starting in the 1970s, neuropsychologists and anthropologists began to realize that other intelligent entities are often the most important parts of the environment to model — because they are the ones modelling you back, whether with friendly or hostile intent. Increasingly intelligent predators put evolutionary pressure on their prey to become smarter, and vice versa."
"The pressures towards intelligence become even more intense for members of social species. Winning mates, sharing resources, gaining followers, teaching, learning and dividing labour: all of these involve modelling and predicting the minds of others. But the more intelligent you become — the better to predict the minds of others (at least in theory) — the more intelligent, and thus hard to predict, those others have also become, because they are of the same species and doing the same thing. These runaway dynamics produce ‘intelligence explosions’. Over the past billion years, symbiogenesis has produced increasingly complex nervous systems, colonies of social animals — and eventually our own technological society. Is this nature’s version of Moore’s law?"
"Since around 2006, transistors have continued to shrink, but the rise in semiconductor operating speed has stalled. To keep increasing computer performance, chip-makers are instead adding more processing cores. They began, in other words, to parallelize silicon-based computation. It’s no coincidence that this is when modern, neural-net-based AI models finally began to take off."
"AI is not distinct from humanity, but rather is a recent addition to a mutually interdependent superhuman entity we are all already part of. An entity that has long been partly biological, partly technological — and always wholly computational. The picture of the future that emerges here is sunnier than that often painted by researchers studying the ethics of AI or its existential risks for humanity. People often presume that evolution — and intelligence — are zero-sum optimization processes, and that AI is both alien to and competitive with humanity. The symbiogenetic view does not guarantee positive outcomes, but neither does it position AI as an alien ‘other’, nor the future as a Malthusian tug-of-war over resources between humans and machines."
John K Clark