War

Hawking and 1000+ other science/tech experts call for ban on AI 'killer robots'

By | | comments |

Weapons that can kill without human operators are 'the third revolution in warfare' artificial intelligence experts warn — and it wouldn't take much to start a global arms race. Nadia Prupis from Common Dreams reports.

More than 1,000 science and technology experts on Monday published an open letter calling for a ban on autonomous weapons – machines capable of killing without human operators – to prevent a "virtually inevitable" high-stakes global arms race.

Among the signatories are physicist and professor Stephen Hawking, Apple co-founder Steve Wozniak, and Tesla CEO Elon Musk, among many others. The letter was presented at the International Joint Conferences on Artificial Intelligence (IJCAI) in Buenos Aires, Argentina.

'Autonomous weapons select and engage targets without human intervention,' the letter states.

It continues:

Artificial Intelligence (AI) technology has reached a point where the deployment of such systems is—practically if not legally—feasible within years, not decades, and the stakes are high: autonomous weapons have been described as the third revolution in warfare, after gunpowder and nuclear arms.'

If any major military power pushes ahead with AI weapon development, a global arms race is virtually inevitable, and the endpoint of this technological trajectory is obvious: autonomous weapons will become the Kalashnikovs of tomorrow.... It will only be a matter of time until they appear on the black market and in the hands of terrorists, dictators wishing to better control their populace, warlords wishing to perpetrate ethnic cleansing, etc. Autonomous weapons are ideal for tasks such as assassinations, destabilizing nations, subduing populations and selectively killing a particular ethnic group.

Proponents of autonomous weapons say the machines would be useful in reducing military casualties on the battlefield. But the letter's authors counter that, in doing so, the weapons would lower the threshold for armed conflict — risking more frequent battles and a greater loss of civilian life.

They state:

'There are many ways in which AI can make battlefields safer for humans, especially civilians, without creating new tools for killing people.'

The letter is the most recent call for a preemptive ban on the weapons, often referred to colloquially as "killer robots".

In April, a joint report by Human Rights Watch and Harvard Law School's International Human Rights Clinic found that autonomous weapons present 'serious moral and legal concerns' and could not only violate international law, but make it virtually impossible to pursue accountability for victims.

The report, titled Mind the Gap: The Lack of Accountability for Killer Robots and presented to the United Nations meeting on lethal weapons, called on the international body to ban such tools before they can be created.

While fully autonomous weapons do not yet exist, HRW noted that their prototypes – such as the U.S. army's Phalanx CIWS and Israel's Iron Dome – are already in use.

This story was originally published in Common Dreams on 27/7/15 under the title 'Hawking, Musk, Woz: We Must Ban the "Kalashnikovs of Tomorrow"' and has been republished under a Creative Commons licence.

Monthly Donation

$

Single Donation

$

Support humanity. Subscribe to IA for just $5.

 
Recent articles by Common Dreams
Three years after January 6, Trump remains grave threat to democracy

In a recent speech to mark the third anniversary of the Capitol attack, President ...  
Chaos erupts after far-Right attack on Brazilian Congress

Far-Right supporters of ousted Brazilian President Jair Bolsonaro have failed to ...  
Elon Musk accused of hypocrisy by kicking critical journalists off Twitter

Musk suspending journalists' accounts is petty and vindictive and absolutely ...  
Join the conversation
comments powered by Disqus

Support Fearless Journalism

If you got something from this article, please consider making a one-off donation to support fearless journalism.

Single Donation

$

Support IAIndependent Australia

Subscribe to IA and investigate Australia today.

Close Subscribe Donate