-
Vietnam, South Korea sign deals on tech, nuclear power
-
EU nears approval of Ukraine loan after Hungary pipeline row
-
Duterte jurisdiction appeal quashed at ICC
-
Three ships targeted in Hormuz, Iran seizes two: monitors, Guards
-
Iran says seized two ships seeking to cross Strait of Hormuz
-
Iran murals project defiance in war with US
-
Oil prices rise despite US-Iran ceasefire extension
-
Ships attacked in Gulf as Trump extends Iran ceasefire
-
Germany set to slash growth forecast due to Mideast war
-
Pakistan's capital holds its breath with US-Iran talks in limbo
-
Groundbreaking Iranian snooker star Vafaei takes on the world
-
Sakib Hussain: IPL quick whose mum sold her jewellery to fund cricket dream
-
US-based Buddhist monks bring peace walk to Sri Lanka
-
NASA unveils new space telescope to give 'atlas of the universe'
-
Trump extends ceasefire, claims Iran 'collapsing financially'
-
The tiny, defiant Nile island caught in the heart of Sudan's war
-
UK inflation jumps as Mideast war propels energy prices
-
Oil falls, stocks mixed as traders weigh outlook after Trump extends truce
-
Oil, stocks mixed as traders weigh outlook after Trump extends truce
-
Anthropic probes unauthorized access to Mythos AI model
-
Stadium that was symbol of NZ post-quake rebuild to hold first match
-
Blazers stun Spurs after Wemby injury, Lakers down Rockets
-
Chinese carmakers aim to build up presence in Europe
-
Maoist landmine legacy haunts India
-
Fiji villagers reject plan for 'Pacific ashtray' in beach paradise
-
India orders school water bells to beat heat
-
Japanese minnows one win from fairytale Champions League title
-
Rugby Australia eyes brighter future as Lions tour brings cash windfall
-
Blazers rally stuns Spurs after Wembanyama injury
-
Young Chinese use AI to launch one-person firms over job anxiety
-
Delicate extraction: Malaysia offers rare earths alternative to China
-
Oil, stocks fall as traders weigh outlook after Trump extends truce
-
Pope to visit prison on final leg of Africa tour
-
US military says key weapons system staying in South Korea
-
India strangles final Maoist bastion as mining looms
-
AI-powered robots offer new hope to German factories
-
Indonesia orangutan forest cleared for 'carbon-neutral' packaging firm
-
PGA Tour mulls pathway back for golfers as LIV plots survival
-
One month phone-free: Young Americans try digital detox
-
Questions about Tesla spending binge ahead of earnings
-
Rome summons Russian ambassador over insults against Meloni
-
US tells Afghans to choose Taliban home or DR Congo: activist
-
John Ternus to lead Apple in the age of AI
-
SpaceX partners with AI startup Cursor, may buy it for $60 bn
-
Mexico pyramid shooter inspired by Columbine attack, pre-Hispanic sacrifices
-
Mexico pyramid shooter planned attack, fixated on US massacre
-
Mbappe on the mark as Real Madrid sink Alaves
-
Rosenior blasts Chelsea flops after 'unacceptable' Brighton defeat
-
Inter roar back to beat Como and reach Italian Cup final
-
Lens sweep past Toulouse to reach French Cup final
AI anxiety as computers get super smart
From Hollywood's death-dealing Terminator to warnings from genius Stephen Hawking or Silicon Valley stars, fears have been fueled that artificial intelligence (AI) could one day destroy humanity.
Tech titans are racing toward creating AI far smarter than people, pushing US President Joe Biden to impose emergency regulation and the European Union seeking major legislation to be agreed by the end of this year.
A two-day summit starting Wednesday in London will explore regulatory safeguards against AI risks such as those below.
- Job stealer? -
The success of ChatGPT from OpenAI has ignited debate about whether "generative AI" capable of quickly producing text, images and audio from simple commands in everyday language is a tremendous threat to jobs held by people.
Automated machinery is already used to do labor in factories, warehouses, and fields.
Generative AI, however, can take aim at white-collar jobs such as lawyers, doctors, teachers, journalists, and even computer programmers.
A report from the McKinsey consulting firm estimates that by the end of this decade, as much as 30 percent of the hours worked in the United States could be automated in a trend accelerated by generative AI.
Boosters of such technology have invoked the notion of a universal basic income in which machines generate wealth that is shared with people freed of the burdens of work.
But it is also possible companies would reap profits of improved efficiencies, leaving those out of work to fend for themselves.
- Copycat? -
Artists were quick to protest software such as Dall-E, Midjourney and Stable Diffusion that are capable of creating images in nearly any style on demand.
Computer coders and writers followed suit, critiquing AI creators for "training" software on their work, enabling it to replicate their styles or skills without permission or compensation.
AI models have been taught using massive amounts of information and imagery found online.
"That's what it trains on, a fraction of the huge output of humanity," OpenAI co-founder Sam Altman said at a conference in September.
"I think this will be a tool that amplifies human beings, not replace them."
- Disinformation tools? -
Fake news and deepfakes have been around for years but being able to easily crank it out using generative AI raises fears of rampant online deception.
Elections run the risk of being won by those most adept at spreading disinformation, contends cognitive scientist and AI expert Gary Marcus.
"Democracy depends on access to the information needed to make the right decisions," Marcus said.
"If no one knows what's true and what's not, it's all over".
- Fraud? -
Generative AI makes it easier for scammers to create convincing phishing emails, perhaps even learning enough about targets to personalize approaches.
Technology lets them copy a face or a voice, and thus trick people into falling for deceptions such as claims a loved one is in danger, for example.
US President Biden called the ability of AI to imitate people's voices "mind blowing" while signing his recent executive order aimed at the technology.
There are even language models trained specifically to produce such malicious content.
- Human role models -
As with other technologies with the potential for good or evil, the main danger is posed by humans who wield it.
Since AI is trained on data put on the web by humans, it can mirror society's prejudices, biases, and injustices.
AI also has the potential to make it easier to create bioweapons; hack banks or power grids; run oppressive government surveillance, and more.
- AI overlord? -
Some industry players fear AI could become so smart that it could seize control from humans.
"It is not difficult to imagine that at some point in the future, our intelligent computers will become as smart or smarter than people," OpenAI co-founder and chief scientist Ilya Sutskever said at a recent TED AI conference.
"The impact of such artificial intelligence is going to be truly vast."
OpenAI and rivals maintain the goal is for AI to benefit humanity, solving long-intractable problems such as climate change.
At the same time, AI industry leaders are calling for thoughtful regulation to prevent risks such as human extinction.
H.Jarrar--SF-PST