Thousands of artificial intelligence (AI) experts and developers have signed a pledge vowing to "neither participate in nor support the development, manufacture, trade, or use of lethal autonomous weapons," and imploring governments worldwide to work together to "create a future with strong international norms, regulations, and laws" barring so-called killer robots.
"We would really like to ensure that the overall impact of the technology is positive and not leading to a terrible arms race, or a dystopian future with robots flying around killing everybody."
More than 160 companies and groups from three dozen countries and 2,400 individuals from 90 countries are backing the pledge, which was developed by the Boston-based Future of Life Institute (FLI) and unveiled Wednesday during the annual International Joint Conference on Artificial Intelligence (IJCAI) in Stockholm, Sweden.
"I'm excited to see AI leaders shifting from talk to action, implementing a policy that politicians have thus far failed to put into effect," declared FLI president and MIT professor Max Tegmark. "AI has huge potential to help the world—if we stigmatize and prevent its abuse. AI weapons that autonomously decide to kill people are as disgusting and destabilizing as bioweapons, and should be dealt with in the same way."
As Anthony Aguirre, a professor at the University of California-Santa Cruz and pledge signatory, told CNN, "We would really like to ensure that the overall impact of the technology is positive and not leading to a terrible arms race, or a dystopian future with robots flying around killing everybody."
Signatory Yoshua Bengio, an AI expert at the Montreal Institute for Learning Algorithms, explained that the pledge has the potential to sway public opinion by shaming developers of killer robots, also referred to called lethal autonomous weapons systems.
SCROLL TO CONTINUE WITH CONTENT
Never Miss a Beat.
Get our best delivered to your inbox.
"This approach actually worked for land mines, thanks to international treaties and public shaming, even though major countries like the U.S. did not sign the treaty banning land mines," Bengio pointed out in an interview with the Guardian. "American companies have stopped building land mines."
Lucy Suchman, a professor at England's Lancaster University, emphasized the importance of AI researchers staying involved with how their inventions are used, noting that as a developer she would, "first, commit to tracking the subsequent uses of my technologies and speaking out against their application to automating target recognition and, second, refuse to participate in either advising or directly helping to incorporate the technology into an autonomous weapon system."
Other high-profile supporters of the pledge include SpaceX and Tesla Motors CEO Elon Musk; Skype founder Jaan Tallinn; Jeffrey Dean, Google's lead of research and machine intelligence; and Demis Hassabis, Shane Legg, and Mustafa Suleyman, the co-founders of DeepMind.
As AI technology has continued to advance, the United Nations has convened a group of governmental experts to address mounting concerns raised by human rights organizations, advocacy groups, military leaders, lawmakers, and tech experts—many who, for years, have demanded a global ban on killer robots.
In recent years, tech experts have used IJCAI as an opportunity to pressure world leaders to outlaw autonomous weapons which, as the new pledge warns, "could become powerful instruments of violence and oppression, especially when linked to surveillance and data systems." Without a ban on such weaponry, they "could too easily spark an arms race that the international community lacks the technical tools and global governance systems to manage."