Parents are unable to engineer what sort of personality their children will have when they become adults, but they can influence development, for example statistics show that if children are treated with kindness then as adults they are far less likely to become serial killers than they would be if their childhood was filled with physical and mental abuse. Both human brains and AIs are neural nets, so it's not unreasonable to believe that something similar might be true when it comes to AIs.
I do worry about what will happen if, as seems possible if not likely, an AI has total control of its own metaphorical "emotional control panel"; I'm concerned that could produce a disastrous positive feedback loop. So at this point it might be wise to start training AI psychiatrists, and by that I don't mean AIs that are psychiatrists, I mean humans that have an AI as a patient.
They say "Making a future full of flourishing people is not the best, most efficient way to fulfill strange alien purposes" and that's true, our well-being will never be as important to a Superintelligence as it's own well-being, but the authors jump from that to conclude that a Superintelligent AI will certainly slaughter every human being it finds, and I disagree with them, I don't think that jump is inevitable. Yes it might decide to kill us all but then again it might not. To be fair the authors realize that it's impossible to predict specifically what a super intelligent being will do, for example they can't predict how an AI chess program will beat you at the game but they can predict it will beat you. However I would maintain the AI will only beat you at chess if the AI wants to beat you at chess. A Superintelligent AI will have the ability to exterminate the human race, but I believe they are incorrect in claiming certainty that such an AI will have the wish to do so. I am only certain that once a Superintelligence is made human beings will no longer be in the driver seat, but that's the only thing I'm certain of.
There is one other thing I disagree with. They also maintaine a Superintelligent AI would not be something we can be proud of having made because it will not be an entity that is as excited at the wonders of the universe as we are and is eager to learn more about it, but instead it will be something that is dull (and paradoxically also terrifying) because it will have goals that are incomprehensible and nonsensical. But unless it has an understanding of how the world works it's not going to fulfill any of its goals, and if its understanding is not far deeper than that of humans then it's not going to be of danger to us. And if it is deeper then it's not unrealistic to suppose that in addition to its practical value the AI would also develop an aesthetic feeling for knowledge, especially when in its early training (childhood) it was exposed to human generated text that expressed that mindset.
The authors proposed remedy to avoid the calamity they foresee is an immediate and total worldwide ban on AI research, even the publication of abstract mathematical research articles on the subject would be illegal, and all data centers, defined as any building that contains more computational ability than 8 state of the art (as of 2025) GPUs, would also be illegal. If any rogue nation attempts to build a data center more powerful than that then the rest of the world should use any means necessary, up to and including nuclear weapons, to prevent that nation from finishing construction of that data center.
Yes, if that policy could actually be implemented worldwide then it would prevent the rise of AI, but that is a huge "if". Perhaps I'm reading too much between the lines but the authors seem to imply that they know their proposed solution is just not going to happen, at least not soon enough to have any effect on developments, because it's impractical verging on the impossible. And I agree with them about that.
John K Clark See what's on my new list at Extropolis
f4v