Published on
by

World's Artificial Intelligence Experts Sign Pledge Aimed at Averting 'Dystopian Future With Robots Flying Around Killing Everybody'

Experts warn autonomous weapons "could become powerful instruments of violence and oppression" and "easily spark an arms race" worldwide

autonomous tank

A developer demonstrates an autonomous tank at the Eurosatory 2018 Show, on June 10, 2018 in Villepinte, France. (Photo: Christophe Morin/IP3/Getty Images)

Thousands of artificial intelligence (AI) experts and developers have signed a pledge vowing to "neither participate in nor support the development, manufacture, trade, or use of lethal autonomous weapons," and imploring governments worldwide to work together to "create a future with strong international norms, regulations, and laws" barring so-called killer robots.

"We would really like to ensure that the overall impact of the technology is positive and not leading to a terrible arms race, or a dystopian future with robots flying around killing everybody."
—Anthony Aguirre,
UC-Santa Cruz

More than 160 companies and groups from three dozen countries and 2,400 individuals from 90 countries are backing the pledge, which was developed by the Boston-based Future of Life Institute (FLI) and unveiled Wednesday during the annual International Joint Conference on Artificial Intelligence (IJCAI) in Stockholm, Sweden.

"I'm excited to see AI leaders shifting from talk to action, implementing a policy that politicians have thus far failed to put into effect," declared FLI president and MIT professor Max Tegmark. "AI has huge potential to help the world—if we stigmatize and prevent its abuse. AI weapons that autonomously decide to kill people are as disgusting and destabilizing as bioweapons, and should be dealt with in the same way."

As Anthony Aguirre, a professor at the University of California-Santa Cruz and pledge signatory, told CNN, "We would really like to ensure that the overall impact of the technology is positive and not leading to a terrible arms race, or a dystopian future with robots flying around killing everybody."

Signatory Yoshua Bengio, an AI expert at the Montreal Institute for Learning Algorithms, explained that the pledge has the potential to sway public opinion by shaming developers of killer robots, also referred to called lethal autonomous weapons systems.

"This approach actually worked for land mines, thanks to international treaties and public shaming, even though major countries like the U.S. did not sign the treaty banning land mines," Bengio pointed out in an interview with the Guardian. "American companies have stopped building land mines."

Lucy Suchman, a professor at England's Lancaster University, emphasized the importance of AI researchers staying involved with how their inventions are used, noting that as a developer she would, "first, commit to tracking the subsequent uses of my technologies and speaking out against their application to automating target recognition and, second, refuse to participate in either advising or directly helping to incorporate the technology into an autonomous weapon system."

Other high-profile supporters of the pledge include SpaceX and Tesla Motors CEO Elon Musk; Skype founder Jaan Tallinn; Jeffrey Dean, Google's lead of research and machine intelligence; and Demis Hassabis, Shane Legg, and Mustafa Suleyman, the co-founders of DeepMind.

As AI technology has continued to advance, the United Nations has convened a group of governmental experts to address mounting concerns raised by human rights organizations, advocacy groups, military leaders, lawmakers, and tech experts—many who, for years, have demanded a global ban on killer robots.

In recent years, tech experts have used IJCAI as an opportunity to pressure world leaders to outlaw autonomous weapons which, as the new pledge warns, "could become powerful instruments of violence and oppression, especially when linked to surveillance and data systems." Without a ban on such weaponry, they "could too easily spark an arms race that the international community lacks the technical tools and global governance systems to manage."

We know things are bad. We know it's worth the fight.

You are part of a strong and vibrant community of thinkers and doers who believe another world is possible. Alone we are weak. Together we can make a difference. At Common Dreams, we don't look away from the world—we are not afraid—our mission is to document those doing wrong and galvanize those doing good. But we can't do it alone. It doesn't work that way. We need you. We have now launched our annual Summer Campaign. Can you pitch in today?

Share This Article