Weapons that can kill without human operators are 'the third revolution in warfare' artificial intelligence experts warn — and it wouldn't take much to start a global arms race. Nadia Prupis from Common Dreams reports.
More than 1,000 science and technology experts on Monday published an open letter calling for a ban on autonomous weapons – machines capable of killing without human operators – to prevent a "virtually inevitable" high-stakes global arms race.
Among the signatories are physicist and professor Stephen Hawking, Apple co-founder Steve Wozniak, and Tesla CEO Elon Musk, among many others. The letter was presented at the International Joint Conferences on Artificial Intelligence (IJCAI) in Buenos Aires, Argentina.
'Autonomous weapons select and engage targets without human intervention,' the letter states.
Artificial Intelligence (AI) technology has reached a point where the deployment of such systems is—practically if not legally—feasible within years, not decades, and the stakes are high: autonomous weapons have been described as the third revolution in warfare, after gunpowder and nuclear arms.'
If any major military power pushes ahead with AI weapon development, a global arms race is virtually inevitable, and the endpoint of this technological trajectory is obvious: autonomous weapons will become the Kalashnikovs of tomorrow.... It will only be a matter of time until they appear on the black market and in the hands of terrorists, dictators wishing to better control their populace, warlords wishing to perpetrate ethnic cleansing, etc. Autonomous weapons are ideal for tasks such as assassinations, destabilizing nations, subduing populations and selectively killing a particular ethnic group.
Stephen Hawking and Elon Musk sign letter suggesting ban on AI warfare http://t.co/7SU3deRbPn pic.twitter.com/NhT6zCcmBN— TechCrunch (@TechCrunch) July 28, 2015
Proponents of autonomous weapons say the machines would be useful in reducing military casualties on the battlefield. But the letter's authors counter that, in doing so, the weapons would lower the threshold for armed conflict — risking more frequent battles and a greater loss of civilian life.
'There are many ways in which AI can make battlefields safer for humans, especially civilians, without creating new tools for killing people.'
The letter is the most recent call for a preemptive ban on the weapons, often referred to colloquially as "killer robots".
In April, a joint report by Human Rights Watch and Harvard Law School's International Human Rights Clinic found that autonomous weapons present 'serious moral and legal concerns' and could not only violate international law, but make it virtually impossible to pursue accountability for victims.
The report, titled Mind the Gap: The Lack of Accountability for Killer Robots and presented to the United Nations meeting on lethal weapons, called on the international body to ban such tools before they can be created.
Mind the gap. The lack of accountability for fully autonomous weapons. New @HRW report http://t.co/t8NXZV2We4 pic.twitter.com/YsFBa8qjRx— Wenzel Michalski (@WenzelMichalski) April 9, 2015
While fully autonomous weapons do not yet exist, HRW noted that their prototypes – such as the U.S. army's Phalanx CIWS and Israel's Iron Dome – are already in use.
This story was originally published in Common Dreams on 27/7/15 under the title 'Hawking, Musk, Woz: We Must Ban the "Kalashnikovs of Tomorrow"' and has been republished under a Creative Commons licence.
Support humanity. Subscribe to IA for just $5.