An AI brain rendered in blue light.

An abstract image of an AI brain.

(Image: Yuichiro Chino/iStock via Getty Images)

If This Is Artificial Intelligence, Let’s Go With Real Stupidity

Six decades after Martin Luther King Jr. warned of “guided missiles and misguided men,” the missiles are even more terrible, the people ordering launches are no less misguided, and the mentalities bent on war are eager to twist AI technology for their own lethal purposes.

At first, I admit, I was a bit flattered to learn that online entrepreneurs are selling study guides for my new book. I thought of CliffsNotes from long ago, helping fellow students who were short on time or interest to grasp the basics of notable works. Curiosity quickly won. I pulled out my credit card, paid $9.99 plus tax for one of the offerings, and awaited its arrival in the mail.

The thin booklet got off to a reasonable enough start, explaining with its first sentence, “The U.S. media coverage that makes it easier to sell wars to the public, as well as the often-hidden cost of civilian casualties from errant U.S. attacks, are all harshly criticized by journalist Solomon.” That wasn’t a bad sum-up of my book.

But the study guide’s second sentence was not nearly as good: “He guarantees that when Russia designated Ukrainian communities during the new attack, the U.S. media was everyone available and jumping into action with compassionate, piercing revealing.” Rereading that sentence a few times didn’t improve it, and I began to worry.

The computer-programmed assaults on the English language escalated. And so, the “war on terror” became the “battle on dread.” The Costs of War project at Brown University became “the Expenses of War project at Earthy Colored College.”

To the extent that meaning could be grasped, the next pages seemed to include some praise: My book “constructs a convincing case that an excessive number of mysteries are being kept from people in general.” What’s more, “the creator presents a sharp and provocative outline of the outcomes of the media’s horrifying disappointments in spreading the word.”

But the study guide also included mild criticism amid the odd wording: “Solomon might have offered a fairly more profound examination of why American newscasting neglects to satisfy its beliefs in covering war and the justifications for why political pioneers could feel a sense of urgency to deal with misdirection while tending to people in general.”

The computer-programmed assaults on the English language escalated. And so, the “war on terror” became the “battle on dread.” A key source of meticulous research that I cited in my book, the Costs of War project at Brown University, became “the Expenses of War project at Earthy Colored College.”

At one point, my book’s actual title—“War Made Invisible”—shifted to “War Caused Imperceptible.” But the laughable malapropisms provided by artificial intelligence became more serious matters when I saw several dozen words forming badly mangled phrases—all attributed to me—inside quotation marks. I could imagine bleary-eyed students cramming on the night before a test or a term-paper deadline, reading the ostensible quotes and thinking that the author of my book must be an idiot.

Likewise, any would-be scholars seeking to glean the gist of the book’s themes in exchange for their $9.99 purchase will surely come away mystified at best after reading sentences like: “It’s totally unsuitable for writers to toe the conflict line for a really long time, and afterward, at last report, essentially, it tends to be informed years past the point of no return.”

I’m not among the authors who claim to never read reviews of their books. In fact, I remember them. So, I could recognize the uber-clumsy efforts of artificial intelligence that sifted through nearly a dozen reviews of War Made Invisible, lifting bits and pieces while weirdly substituting supposed synonyms to steer clear of plagiarism lawsuits.

So, let’s hear it for digital “free enterprise.” Or maybe that’s “unshackled business.” Nice AI work if you can get it.

Which brings us to a vastly more substantive matter. Artificial so-called intelligence is hardly immune to a dynamic that computer experts long ago dubbed “GIGO”—garbage in, garbage out. With AI, no matter how sophisticated it might seem, the consequences in war are apt to be horrific. Six decades after Martin Luther King Jr. warned of “guided missiles and misguided men,” the missiles are even more terrible, the people ordering launches are no less misguided, and the mentalities bent on war are eager to twist AI technology for their own lethal purposes.

A couple of weeks ago, the Department of Defense (DOD) announced “the establishment of a generative artificial intelligence task force, an initiative that reflects the DOD’s commitment to harnessing the power of artificial intelligence in a responsible and strategic manner.”

If they were still alive, the 4.5 million people who have died as direct and indirect results of U.S. wars since 9/11 might doubt how “responsible” the Defense Department’s manner has been.

Let’s hope that the people running the Pentagon’s task force for artificial intelligence didn’t graduate from Earthy Colored College.

Join Us: News for people demanding a better world


Common Dreams is powered by optimists who believe in the power of informed and engaged citizens to ignite and enact change to make the world a better place.

We're hundreds of thousands strong, but every single supporter makes the difference.

Your contribution supports this bold media model—free, independent, and dedicated to reporting the facts every day. Stand with us in the fight for economic equality, social justice, human rights, and a more sustainable future. As a people-powered nonprofit news outlet, we cover the issues the corporate media never will. Join with us today!

Our work is licensed under Creative Commons (CC BY-NC-ND 3.0). Feel free to republish and share widely.