This little ditty in Science News, Our Final Invention, about the impending dangers of Artificial Intelligence caught my eye. Consider this quote, “Computer systems advanced enough to act with human-level intelligence will likely be unpredictable and inscrutable all of the time.” He is worried that, “we could end up with a planet self-serving, self-replicating AI entities that act ruthlessly toward their creators”.
Apparently, an increasing number of AI visionaries also have misgivings. Such “The sky is falling” fears amount to blindness through vision, as I see it. If history shows us anything, it is that the pageant of history invariably turns out differently from that of our desires or worries, in the long-term. Indeed, history is a record of humanity being blindsided at every turn.
Their mistake is to equate real intelligence with the human ability to think. They fear AI will out-think and outwit humans. Such so-called intelligence is just the tip of the iceberg. These visionaries fail to realize that deep seated emotion drives thinking. As such, thinking can be no more ‘intelligent’ than one’s underlying emotional state. Emotion is the foundational driving force for what all animals (including humans) do in life. For us, emotion steers thinking, often imperceptibly. Why haven’t I heard any mention of ‘AE’, i.e., artificial emotion? Any nefarious agendas we fear originate in emotion, not in intelligence. What ‘disease’ causes us to attribute excessive significance to intelligence?
I can’t help but see how we are blinded by our own self-image. We see intelligence as key, and often see emotion as something to ‘rise above’. What we fail to realize is that any ‘rising above’ must also be driven by emotion. Emotion is the fuel that runs life. Intelligence is merely a species-specific neurological ability to get the task done—the task that instinct and emotion dictate. For birds, that means flying south for the winter; for humans that means cognition. What sets humans apart from the rest is our ability to imagine—to fly around in ‘mind space’. Still, that imagination is still steered by core emotion: fear and need. (See Need and Fear in One who speaks does not know?).
Our species’ illusion-of-self (ego) drives our cherished views of intelligence, probably more now than at anytime in the past (1). Cognition is to us what a dexterous long truck is to an elephant. The difference lies in how intelligence enables us to focus on intelligence to the point of becoming stupidly intelligent. It would be like the elephant evolving the length of its trunk to the point that it began tripping over it. Our intelligence has become something of a disease, as chapter 71 points out: Realizing I don’t’ know is better; not knowing this knowing is disease. D.C. Lau put this more diplomatically as, Not to know yet to think that one knows will lead to difficulty. Here now is the rest of this review, interspersed with my snarky remarks when I can’t resist. 😉
Computers already make all sorts of decisions for you. With little or no human guidance, they deduce what books you would like to buy, trade your stocks and distribute electrical power. They do all this quickly and efficiently using a simple form of artificial intelligence. Now, imagine if computers controlled even more aspects of life and could truly think for themselves.
What about computers feeling for themselves? If a computer can’t care, need or fear for itself, what motivates it to ‘think or act’ for itself. In other words, let’s not forget the prerequisites for an ‘illusion of self’ necessary for computers to “think for themselves”. (See Buddha’s Second Noble Truth.)
Barrat, a documentary filmmaker and author, chronicles his discussions with scientists and engineers who are developing ever more complex artificial intelligence, or AI. The goal of many in the field is to make a mechanical brain as intelligent — creative, flexible and capable of learning — as the human mind. But an increasing number of AI visionaries have misgivings.
Science fiction has long explored the implications of humanlike machines (think of Asimov’s I, Robot), but Barrat’s thoughtful treatment adds a dose of reality. Through his conversations with experts, he argues that the perils of AI can easily, even inevitably, outweigh its promise.
Certainly, the perils of AI “can easily outweigh their promise”, at least in the short term. It is already true of computers now in many ways. Hey, it is even true of our natural intelligence, i.e., the ‘not knowing this knowing disease’ to which chapter 71 refers. However, evolution is like a child in how it stumbles along, striving until it gets it right, and then ups-the-anti, and stumbles along at that new level… ad infinitum. We’re just part of that cosmic process, and the invention of computers is just the latest in our evolutionary upping-the-anti.
By mid-century — maybe within a decade, some researchers say — a computer may achieve human-scale artificial intelligence, an admittedly fuzzy milestone. (The Turing test provides one definition: a computer would pass the test by fooling humans into thinking it’s human.) AI could then quickly evolve to the point where it is thousands of times smarter than a human. But long before that, an AI robot or computer would become self-aware and would not be interested in remaining under human control, Barrat argues.
Ha! “A thousand times smarter than a human, self-aware, and not interested…”. Smarter by whose definition of what smart actually is. As to computers being “self-aware”, Buddha’s “the illusion of self originates and manifests itself in a cleaving to things” demolishes that notion. The need to “cleave” is a pure, raw, visceral emotion. Computers don’t feel need or fear, and I’ve yet to see any attempt to create an artificial version of these bio-chemical based emotions. Without that, the whole doomsday story falls apart.
One AI researcher notes that self-aware, self-improving systems will have three motivations: efficiency, self-protection and acquisition of resources, primarily energy. Some people hesitate to even acknowledge the possible perils of this situation, believing that computers programmed to be superintelligent can also be programmed to be “friendly.” But others, including Barrat, fear that humans and AI are headed toward a mortal struggle. Intelligence isn’t unpredictable merely some of the time or in special cases, he writes. “Computer systems advanced enough to act with human-level intelligence will likely be unpredictable and inscrutable all of the time.”
The researcher noted, “…systems will have three motivations: efficiency, self-protection and acquisition of resources, primarily energy”. Okay, I’m probably beating a dead horse by now. But really, “self-protection”? Yes, you can program a computer to protect itself from threats that the programmer stipulates. Those are merely projections of self-survival of the programmer. They originate in his survival instinct — ergo, his agenda. I’ve heard no mention of anyone designing Artificial Instinct.
Each child becomes an adult, yet the child remains within the adult. Most, if not all, the havoc we wreak stems from ignorance, not intelligence, or rather a mismatch between emotional intelligence and cognitive intelligence. (See, Counterbalancing I.Q.). Therefore, for AI to wreak havoc it must also possess Artificial Ignorance.
Humans, he says, need to figure out now, at the early stages of AI’s creation, how to coexist with hyperintelligent machines. Otherwise, Barrat worries, we could end up with a planet — eventually a galaxy — populated by self-serving, self-replicating AI entities that act ruthlessly toward their creators.
That just goes to show that we will always find something to worry about. Simply put, worry = fear + thought and this will always find a way to express itself as long as we feel fear and we think. The wealthier we become, the more absurd and neurotic those worries will end up. Jesus saw this consequence when he said, “It is easier for a camel to go through the eye of a needle, than for a rich man to enter into the kingdom of God.” This “Kingdom of God” for me is simply a state of natural dynamic balance. Wealth skews balance. The harnessing of electricity has made us all wealthy, relatively speaking. (See, And Then There Was Fire for an overview of our evolutionary progress to date, and a glimpse of a hopeful long-term future that we are stumbling toward.)
All this fuss is merely another example of instinct treading water to find meaning. No one is to blame… not one bit! After all, without free will, the cards just fall where they may. For more on this worry = fear + thought, and Emotional Intelligence, see Counterbalancing I.Q., Beware: the Blind Spot, and Imagining a Better Way.
(1) Searching for the Mind (jonlieffmd.com) exemplifies the importance we play on intelligence. We see intelligence as strength; do we also sense it as weakness? Two sides of one coin. Anyway, investigating this scientifically is bound to make us more self-honest.