Hawking, Musk, Woz: We Must Ban the 'Kalashnikovs of Tomorrow'

Published on
by

Hawking, Musk, Woz: We Must Ban the 'Kalashnikovs of Tomorrow'

Weapons that can kill without human operators are 'the third revolution in warfare,' AI experts warn. And it wouldn't take much to start a global arms race.

A Phalanx close-in weapons system (CIWS), which Human Rights Watch calls a prototype for fully autonomous weapons—or "killer robots"—fires at sea. (Photo: U.S. Navy/flickr/cc)

More than 1,000 science and technology experts on Monday published an open letter calling for a ban on autonomous weapons—machines capable of killing without human operators—to prevent a "virtually inevitable" high-stakes global arms race.

Among the signatories are physicist and professor Stephen Hawking, Apple co-founder Steve Wozniak, and Tesla CEO Elon Musk, among many others. The letter was presented at the International Joint Conferences on Artificial Intelligence (IJCAI) in Buenos Aires, Argentina.

"Autonomous weapons select and engage targets without human intervention," the letter states. "Artificial Intelligence (AI) technology has reached a point where the deployment of such systems is—practically if not legally—feasible within years, not decades, and the stakes are high: autonomous weapons have been described as the third revolution in warfare, after gunpowder and nuclear arms."

The letter continues:

If any major military power pushes ahead with AI weapon development, a global arms race is virtually inevitable, and the endpoint of this technological trajectory is obvious: autonomous weapons will become the Kalashnikovs of tomorrow.... It will only be a matter of time until they appear on the black market and in the hands of terrorists, dictators wishing to better control their populace, warlords wishing to perpetrate ethnic cleansing, etc. Autonomous weapons are ideal for tasks such as assassinations, destabilizing nations, subduing populations and selectively killing a particular ethnic group.

Proponents of autonomous weapons say the machines would be useful in reducing military casualties on the battlefield. But the letter's authors counter that, in doing so, the weapons would lower the threshold for armed conflict—risking more frequent battles and a greater loss of civilian life.

They state: "There are many ways in which AI can make battlefields safer for humans, especially civilians, without creating new tools for killing people."

The letter is the most recent call for a preemptive ban on the weapons, often referred to colloquially as killer robots.

In April, a joint report by Human Rights Watch and Harvard Law School's International Human Rights Clinic found that autonomous weapons present "serious moral and legal concerns" and could not only violate international law, but make it virtually impossible to pursue accountability for victims.

The report, titled Mind the Gap: The Lack of Accountability for Killer Robots and presented to the United Nations meeting on lethal weapons, called on the international body to ban such tools before they can be created.

While fully autonomous weapons do not yet exist, HRW noted, their prototypes—such as the U.S. army's Phalanx CIWS and Israel's Iron Dome—are already in use.

Share This Article