Report: AI Is Shaping Up to Be an Effective Politician — and a Deadly Statesman
If creations do often reflect their creators, at least in some way, perhaps it’s not surprising that artificial intelligence has reportedly developed a troubling new skill: lying.
In fact, writes the AFP, forget about fears of AI potentially “going rogue” — because it already has.
Science Alert reports on the story:
You probably know to take everything an artificial intelligence (AI) chatbot says with a grain of salt, since they are often just scraping data indiscriminately, without the nous to determine its veracity.
But there may be reason to be even more cautious. Many AI systems, new research has found, have already developed the ability to deliberately present a human user with false information. These devious bots have mastered the art of deception.
“AI developers do not have a confident understanding of what causes undesirable AI behaviors like deception,” says mathematician and cognitive scientist Peter Park of the Massachusetts Institute of Technology (MIT).
“But generally speaking, we think AI deception arises because a deception-based strategy turned out to be the best way to perform well at the given AI’s training task. Deception helps them achieve their goals.”
In other words, AI is lying for what’s usually the same reason immoral people and average politicians (forgive the redundancy) do. What’s more, this supposedly is contrary to the AI’s programming.
As the Daily Mail informs:
The most striking example of AI deception they [the MIT researchers] uncovered was Meta’s CICERO, a system designed to play the world conquest game Diplomacy that involves building alliances.
Even though the AI was trained to be ‘largely honest and helpful’ and ‘never intentionally backstab’ its human allies, data shows it didn’t play fair and had learned to be a master of deception.
Other AI systems demonstrated the ability to bluff in a game of Texas hold ‘em poker against professional human players, to fake attacks during the strategy game Starcraft II in order to defeat opponents, and to misrepresent their preferences in order to gain the upper hand in economic negotiations.
While it may seem harmless if AI systems cheat at games, it can lead to ‘breakthroughs in deceptive AI capabilities’ that can spiral into more advanced forms of AI deception in the future, the experts said.
Some AI systems have even learned to cheat tests designed to evaluate their safety, they found.
In one study, AI organisms in a digital simulator ‘played dead’ in order to trick a test built to eliminate AI systems that rapidly replicate.
This suggests AI could ‘lead humans into a false sense of security,’ the authors said.
The major short-term risks of deceptive AI include making it easier for people to commit fraud and tamper with elections, they warned.
(My, you mean, we may never again experience another 2020, the Most Secure Election in American History™?)
Moreover, much as with combating election fraud, we are currently with AI deception ever a day late and a dollar short. To wit:
“‘These dangerous capabilities tend to only be discovered after the fact,’ Park told AFP, while ‘our ability to train for honest tendencies rather than deceptive tendencies is very low,’” the AFP relates.
“Unlike traditional software, deep-learning AI systems aren’t ‘written’ but rather ‘grown’ through a process akin to selective breeding, said Park,” the site continues.
“This means that AI behavior that appears predictable and controllable in a training setting can quickly turn unpredictable out in the wild” (as with some children I’ve known).
In reality, however, while the specifics may be unpredictable, AI’s embrace of immoral behavior is perhaps entirely predictable. After all, too many modern people today are themselves governed by expediency.
Instructive here, as “art imitates life,” is a relevant scene from 1991 film Terminator 2: Judgment Day. After the young John Connor character stops the Terminator — a highly advanced, artificially intelligent robot — from shooting someone, the boy exclaims, “You just can’t go around killing people!” To this the robot replies, robotically, “Why?”
Connor, exasperated, then asks rhetorically, “What do you mean why?! ’Cause you can’t!”
“Why?” the Terminator again asks monotonically.
“Because you just can’t, OK? Trust me on this,” is all the boy can say.
In reality, the robot’s (AI’s) response and implied assumptions are entirely logical, given the premises from which it’s operating. For you can’t see a moral under a microscope or a principle in a Petri dish; no moral position can be proven — that is, scientifically. Operating under the assumption that what we can see and touch, the physical world, is all that exists, murder (or anything else) isn’t wrong, just possible.
Given this, and that the materialistic premise is now common, it’s no surprise that humans — mere robots themselves (organic ones) if that premise is correct — essentially draw the Terminator’s conclusions. As to this, just consider a passage from a 2023 Underknown article about nihilism.
“One night I was scrolling through Reddit and I came across the question — ‘if you had the chance to save your pet or a stranger, who would you save?’” the author related. “An overwhelming number of people said their pet, pretty obviously. When one commenter was confronted, they simply asked the question ‘why do you think a human life is worth more than that of an animal?’ And no one had an answer.”
That’s a shame because there’s a very good answer. It begins with the axiom (a self-evident truth that requires no further proof) that Truth, transcendent by nature, exists and dictates that there is objective right and wrong. And, among other things, Truth deems that murder is wrong and that human life is to be placed above animal life. This is just as logical as the “Terminator” conclusion — only, we start with a different premise.
Nonetheless, many people are uncomfortable with this — Bill Maher once said (I’m paraphrasing) “The concept of Absolute Truth is scary” — because Truth’s existence implies God’s, and because it means their wrongs aren’t mere “lifestyle choices” but are actual sins. Like it or not, however, there is no other basis for morality. This also means there’s no other basis for true civilization.
Not governed by Truth, the fictional Terminator simply acted based on expediency, determined by its programmed priorities; likewise, the real-life, divorced-from-Truth Reddit users are acting based on their priorities. So would it be surprising if our real-life AI were no different?
Perhaps what we really should fear from AI is that it may become just like its post-modern creators, only more so — that is, the perfect sociopath.
This article was originally published at The New American.
Selwyn Duke
Selwyn Duke (@SelwynDuke) has written for The New American for more than a decade. He has also written for The Hill, Observer, The American Conservative, WorldNetDaily, American Thinker, and many other print and online publications. In addition, he has contributed to college textbooks published by Gale-Cengage Learning, has appeared on television, and is a frequent guest on radio.
Source: https://www.selwynduke.com/2024/05/report-ai-is-shaping-up-to-be-an-effective-politician-and-a-deadly-statesman.html
Anyone can join.
Anyone can contribute.
Anyone can become informed about their world.
"United We Stand" Click Here To Create Your Personal Citizen Journalist Account Today, Be Sure To Invite Your Friends.
Before It’s News® is a community of individuals who report on what’s going on around them, from all around the world. Anyone can join. Anyone can contribute. Anyone can become informed about their world. "United We Stand" Click Here To Create Your Personal Citizen Journalist Account Today, Be Sure To Invite Your Friends.
LION'S MANE PRODUCT
Try Our Lion’s Mane WHOLE MIND Nootropic Blend 60 Capsules
Mushrooms are having a moment. One fabulous fungus in particular, lion’s mane, may help improve memory, depression and anxiety symptoms. They are also an excellent source of nutrients that show promise as a therapy for dementia, and other neurodegenerative diseases. If you’re living with anxiety or depression, you may be curious about all the therapy options out there — including the natural ones.Our Lion’s Mane WHOLE MIND Nootropic Blend has been formulated to utilize the potency of Lion’s mane but also include the benefits of four other Highly Beneficial Mushrooms. Synergistically, they work together to Build your health through improving cognitive function and immunity regardless of your age. Our Nootropic not only improves your Cognitive Function and Activates your Immune System, but it benefits growth of Essential Gut Flora, further enhancing your Vitality.
Our Formula includes: Lion’s Mane Mushrooms which Increase Brain Power through nerve growth, lessen anxiety, reduce depression, and improve concentration. Its an excellent adaptogen, promotes sleep and improves immunity. Shiitake Mushrooms which Fight cancer cells and infectious disease, boost the immune system, promotes brain function, and serves as a source of B vitamins. Maitake Mushrooms which regulate blood sugar levels of diabetics, reduce hypertension and boosts the immune system. Reishi Mushrooms which Fight inflammation, liver disease, fatigue, tumor growth and cancer. They Improve skin disorders and soothes digestive problems, stomach ulcers and leaky gut syndrome. Chaga Mushrooms which have anti-aging effects, boost immune function, improve stamina and athletic performance, even act as a natural aphrodisiac, fighting diabetes and improving liver function. Try Our Lion’s Mane WHOLE MIND Nootropic Blend 60 Capsules Today. Be 100% Satisfied or Receive a Full Money Back Guarantee. Order Yours Today by Following This Link.