Published on

World's Artificial Intelligence Experts Sign Pledge Aimed at Averting 'Dystopian Future With Robots Flying Around Killing Everybody'

Experts warn autonomous weapons "could become powerful instruments of violence and oppression" and "easily spark an arms race" worldwide

autonomous tank

A developer demonstrates an autonomous tank at the Eurosatory 2018 Show, on June 10, 2018 in Villepinte, France. (Photo: Christophe Morin/IP3/Getty Images)

Thousands of artificial intelligence (AI) experts and developers have signed a pledge vowing to "neither participate in nor support the development, manufacture, trade, or use of lethal autonomous weapons," and imploring governments worldwide to work together to "create a future with strong international norms, regulations, and laws" barring so-called killer robots.

"We would really like to ensure that the overall impact of the technology is positive and not leading to a terrible arms race, or a dystopian future with robots flying around killing everybody."
—Anthony Aguirre,
UC-Santa Cruz

More than 160 companies and groups from three dozen countries and 2,400 individuals from 90 countries are backing the pledge, which was developed by the Boston-based Future of Life Institute (FLI) and unveiled Wednesday during the annual International Joint Conference on Artificial Intelligence (IJCAI) in Stockholm, Sweden.

"I'm excited to see AI leaders shifting from talk to action, implementing a policy that politicians have thus far failed to put into effect," declared FLI president and MIT professor Max Tegmark. "AI has huge potential to help the world—if we stigmatize and prevent its abuse. AI weapons that autonomously decide to kill people are as disgusting and destabilizing as bioweapons, and should be dealt with in the same way."

As Anthony Aguirre, a professor at the University of California-Santa Cruz and pledge signatory, told CNN, "We would really like to ensure that the overall impact of the technology is positive and not leading to a terrible arms race, or a dystopian future with robots flying around killing everybody."

Signatory Yoshua Bengio, an AI expert at the Montreal Institute for Learning Algorithms, explained that the pledge has the potential to sway public opinion by shaming developers of killer robots, also referred to called lethal autonomous weapons systems.


Never Miss a Beat.

Get our best delivered to your inbox.

"This approach actually worked for land mines, thanks to international treaties and public shaming, even though major countries like the U.S. did not sign the treaty banning land mines," Bengio pointed out in an interview with the Guardian. "American companies have stopped building land mines."

Lucy Suchman, a professor at England's Lancaster University, emphasized the importance of AI researchers staying involved with how their inventions are used, noting that as a developer she would, "first, commit to tracking the subsequent uses of my technologies and speaking out against their application to automating target recognition and, second, refuse to participate in either advising or directly helping to incorporate the technology into an autonomous weapon system."

Other high-profile supporters of the pledge include SpaceX and Tesla Motors CEO Elon Musk; Skype founder Jaan Tallinn; Jeffrey Dean, Google's lead of research and machine intelligence; and Demis Hassabis, Shane Legg, and Mustafa Suleyman, the co-founders of DeepMind.

As AI technology has continued to advance, the United Nations has convened a group of governmental experts to address mounting concerns raised by human rights organizations, advocacy groups, military leaders, lawmakers, and tech experts—many who, for years, have demanded a global ban on killer robots.

In recent years, tech experts have used IJCAI as an opportunity to pressure world leaders to outlaw autonomous weapons which, as the new pledge warns, "could become powerful instruments of violence and oppression, especially when linked to surveillance and data systems." Without a ban on such weaponry, they "could too easily spark an arms race that the international community lacks the technical tools and global governance systems to manage."

This is the world we live in. This is the world we cover.

Because of people like you, another world is possible. There are many battles to be won, but we will battle them together—all of us. Common Dreams is not your normal news site. We don't survive on clicks. We don't want advertising dollars. We want the world to be a better place. But we can't do it alone. It doesn't work that way. We need you. If you can help today—because every gift of every size matters—please do. Without Your Support We Simply Don't Exist.

Please select a donation method:

Share This Article