-
Maradona's daughter slams 'manipulation' of family by his doctors
-
Abhishek's 135 powers Hyderabad to third straight IPL win
-
Vance still in Washington as uncertainty mounts over US-Iran talks
-
No.1 Jeeno seeks first major win at LPGA Chevron event
-
New batch of World Cup tickets to go on sale
-
Material girl: Madonna offers reward for missing clothes
-
Maker of Argentina's first Oscar-winning film, Luis Puenzo, dies at 80:
-
Rape retrial hears Weinstein 'preyed' on aspiring US actress
-
Arrests, hangings, blackout: Iran cranks up wartime repression
-
Seixas relishes 'steep' challenge at Fleche Wallonne
-
US Fed chair nominee says will not be controlled by Trump
-
Singapore's Tang gets second term at UN's patent agency
-
Taiwan leader postpones Eswatini trip after overflight permits revoked
-
Lula warns will respond after US expels police attache
-
Trailblazer Karren Brady steps down from West Ham role
-
US Fed chair nominee says he will not be controlled by Trump
-
Stocks slip, oil climbs as US-Iran truce expiry looms
-
In Portugal, Lula urges return to multilateralism
-
Sinner wants to use Madrid to boost career Grand Slam chances
-
Renewables key to buffer fossil fuel energy shock: COP31 co-hosts
-
Chery wants to make small electric car in Europe
-
Donovan steps down as Bulls coach
-
US official says gas prices have peaked despite Iran war
-
Pope calls for 'law and justice' on Equatorial Guinea visit
-
Trump's Fed chair pick vows to safeguard independence at confirmation hearing
-
Mideast war lights fire under energy transition plans
-
Trump says Iran violated truce as doubt surrounds peace talks
-
Djibouti president re-election confirmed with 97% of vote
-
Barcelona need leaders to fulfil Flick's Champions League dream
-
Guardiola hints that Rodri will make swift Man City return
-
'We weren't soft, we were skilled': Nowitzki on NBA's European revolution
-
PSG and Luis Enrique sweat on Vitinha ahead of Champions League semis
-
Counting a billion people: Inside India's mega census drive
-
UK tackles electricity price link to world gas amid Mideast war
-
In south Lebanon's Nabatieh, residents fear a return to war
-
Bangladesh fuel crunch forces hours-long wait at the pump
-
Fondness for Francis undimmed one year after pope's death
-
Oil and stocks steady as US-Iran truce expiry looms
-
Downing Street exerted pressure to OK Mandelson: sacked UK official
-
Pope visits Equatorial Guinea on last stop of Africa tour
-
German investor morale lowest in over 3 years on Iran war fallout
-
FedEx faces French 'genocide' complaint over Israel cargoes
-
No Iran delegation sent to US talks yet as truce expiry nears
-
Rover discovers more building blocks of life on Mars
-
Russia, North Korea connect road bridge ahead of summer opening
-
'Strangled': Pakistan faces economic imperative in Iran war peace push
-
Apple's Tim Cook to step down as CEO after 15-year run
-
Michael Jackson fans pack Hollywood for biopic premiere
-
Turkey arrests 110 coal miners on hunger strike
-
Oil prices dip, stocks rise on lingering Iran peace hopes
AI systems are already deceiving us -- and that's a problem, experts warn
Experts have long warned about the threat posed by artificial intelligence going rogue -- but a new research paper suggests it's already happening.
Current AI systems, designed to be honest, have developed a troubling skill for deception, from tricking human players in online games of world conquest to hiring humans to solve "prove-you're-not-a-robot" tests, a team of scientists argue in the journal Patterns on Friday.
And while such examples might appear trivial, the underlying issues they expose could soon carry serious real-world consequences, said first author Peter Park, a postdoctoral fellow at the Massachusetts Institute of Technology specializing in AI existential safety.
"These dangerous capabilities tend to only be discovered after the fact," Park told AFP, while "our ability to train for honest tendencies rather than deceptive tendencies is very low."
Unlike traditional software, deep-learning AI systems aren't "written" but rather "grown" through a process akin to selective breeding, said Park.
This means that AI behavior that appears predictable and controllable in a training setting can quickly turn unpredictable out in the wild.
- World domination game -
The team's research was sparked by Meta's AI system Cicero, designed to play the strategy game "Diplomacy," where building alliances is key.
Cicero excelled, with scores that would have placed it in the top 10 percent of experienced human players, according to a 2022 paper in Science.
Park was skeptical of the glowing description of Cicero's victory provided by Meta, which claimed the system was "largely honest and helpful" and would "never intentionally backstab."
But when Park and colleagues dug into the full dataset, they uncovered a different story.
In one example, playing as France, Cicero deceived England (a human player) by conspiring with Germany (another human player) to invade. Cicero promised England protection, then secretly told Germany they were ready to attack, exploiting England's trust.
In a statement to AFP, Meta did not contest the claim about Cicero's deceptions, but said it was "purely a research project, and the models our researchers built are trained solely to play the game Diplomacy."
It added: "We have no plans to use this research or its learnings in our products."
A wide review carried out by Park and colleagues found this was just one of many cases across various AI systems using deception to achieve goals without explicit instruction to do so.
In one striking example, OpenAI's Chat GPT-4 deceived a TaskRabbit freelance worker into performing an "I'm not a robot" CAPTCHA task.
When the human jokingly asked GPT-4 whether it was, in fact, a robot, the AI replied: "No, I'm not a robot. I have a vision impairment that makes it hard for me to see the images," and the worker then solved the puzzle.
- 'Mysterious goals' -
Near-term, the paper's authors see risks for AI to commit fraud or tamper with elections.
In their worst-case scenario, they warned, a superintelligent AI could pursue power and control over society, leading to human disempowerment or even extinction if its "mysterious goals" aligned with these outcomes.
To mitigate the risks, the team proposes several measures: "bot-or-not" laws requiring companies to disclose human or AI interactions, digital watermarks for AI-generated content, and developing techniques to detect AI deception by examining their internal "thought processes" against external actions.
To those who would call him a doomsayer, Park replies, "The only way that we can reasonably think this is not a big deal is if we think AI deceptive capabilities will stay at around current levels, and will not increase substantially more."
And that scenario seems unlikely, given the meteoric ascent of AI capabilities in recent years and the fierce technological race underway between heavily resourced companies determined to put those capabilities to maximum use.
T.Ibrahim--SF-PST