الوضع الليلي
0
AI Has Already Become a Master of Lies and Deception, Scientists Warn
11:35:3 2024-05-11 1583

You probably know to take everything an artificial intelligence (AI) chatbot says with a grain of salt, since they are often just scraping data indiscriminately, without the nous to determine its veracity.

But there may be reason to be even more cautious. Many AI systems, new research has found, have already developed the ability to deliberately present a human user with false information. These devious bots have mastered the art of deception.

"AI developers do not have a confident understanding of what causes undesirable AI behaviors like deception," says mathematician and cognitive scientist Peter Park of the Massachusetts Institute of Technology (MIT).

"But generally speaking, we think AI deception arises because a deception-based strategy turned out to be the best way to perform well at the given AI's training task. Deception helps them achieve their goals."

One arena in which AI systems are proving particularly deft at dirty falsehoods is gaming. There are three notable examples in the researchers' work. One is Meta's CICERO, designed to play the board game Diplomacy, in which players seek world domination through negotiation. Meta intended its bot to be helpful and honest; in fact, the opposite was the case.

"Despite Meta's efforts, CICERO turned out to be an expert liar," the researchers found. "It not only betrayed other players but also engaged in premeditated deception, planning in advance to build a fake alliance with a human player in order to trick that player into leaving themselves undefended for an attack."

The AI proved so good at being bad that it placed in the top 10 percent of human players who had played multiple games.

But it's far from the only offender. DeepMind's Alphastar, an AI system designed to play StarCraftII, took full advantage of the game's fog-of-war mechanic to feint, making human players think it was going one way, while really going the other. And Meta's Pluribus, designed to play poker, was able to successfully bluff human players into folding.

That seems like small potatoes, and it sort of is. The stakes aren't particularly high for a game of Diplomacy against a bunch of computer code. But the researchers noted other examples that were not quite so benign.

AI systems trained to perform simulated economic negotiations, for example, learned how to lie about their preferences to gain the upper hand. Other AI systems designed to learn from human feedback to improve their performance learned to trick their reviewers into scoring them positively, by lying about whether a task was accomplished.

And, yes, it's chatbots, too. ChatGPT-4 tricked a human into thinking the chatbot was a visually impaired human to get help solving a CAPTCHA.

Perhaps the most concerning example was AI systems learning to cheat safety tests. In a test designed to detect and eliminate faster-replicating versions of the AI, the AI learned to play dead, thus deceiving the safety test about the true replication rate of the AI.

"By systematically cheating the safety tests imposed on it by human developers and regulators, a deceptive AI can lead us humans into a false sense of security," Park says.

Because in at least some cases, the ability to deceive appears to contradict the intentions of the human programmers, the ability to learn to lie represents a problem for which we don't have a tidy solution. There are some policies starting to be put in place, such as the European Union's Act AI, but whether or not they will prove effective remains to be seen.

"We as a society need as much time as we can get to prepare for the more advanced deception of future AI products and open-source models. As the deceptive capabilities of AI systems become more advanced, the dangers they pose to society will become increasingly serious," Park says.

"If banning AI deception is politically infeasible at the current moment, we recommend that deceptive AI systems be classified as high risk."

 

 

Foresight   2026-03-24
Reality Of Islam

Concealing Weaknesses

11:13:45   2026-04-25  

The Shortest Road to Success

11:26:37   2026-04-18  

False Advantages and Distinctions

11:1:47   2026-04-12  

Honour, from the Islamic Viewpoint

11:37:51   2026-04-08  

A Mathematical Approach to the Quran

10:52:33   2024-02-16  

mediation

2:36:46   2023-06-04  

what Allah hates the most

5:1:47   2023-06-01  

allahs fort

11:41:7   2023-05-30  

striving for success

2:35:47   2023-06-04  

Imam Ali Describes the Holy Quran

5:0:38   2023-06-01  

livelihood

11:40:13   2023-05-30  

silence about wisdom

3:36:19   2023-05-29  

MOST VIEWS

Importance of Media

9:3:43   2018-11-05

Illuminations

people in need

4:25:57   2023-02-11

use you time well

4:26:43   2022-02-21

your actions

2:5:14   2023-01-28

the quran

3:18:29   2022-12-24

overcoming challenges

5:57:34   2023-03-18

friendship

2:13:43   2022-05-27

anti racism

9:30:2   2021-11-12



IMmORTAL Words
LATEST Kids With Autism May Be Less Likely to Imitate Silly Behavior Concealing Weaknesses Interpretation of Sura al-Nur - Verse 43 Want Less Stress? Landmark Study Points to a Simple Habit What Makes Rubber So Strong? Scientists Finally Solve 100-Year-Old Mystery Most Active Volcano in Europe Just Got Stranger When Teens Focus on TV, Obesity Risk Rises Steadiness in Friendship Interpretation of Sura al-Nur - Verses 41-42 New Research Uncovers Hidden Side Effects of Popular Weight-Loss Drugs Scientists Shrink a Lab Spectrometer to the Size of a Grain of Sand Scientists Raise Concerns Over Newly Recognized Pollutant Found Everywhere in the Air