

SUBSCRIBE TO OUR FREE NEWSLETTER
Daily news & progressive opinion—funded by the people, not the corporations—delivered straight to your inbox.
5
#000000
#FFFFFF
To donate by check, phone, or other method, see our More Ways to Give page.


Daily news & progressive opinion—funded by the people, not the corporations—delivered straight to your inbox.

An abstract image of an AI brain.
Six decades after Martin Luther King Jr. warned of “guided missiles and misguided men,” the missiles are even more terrible, the people ordering launches are no less misguided, and the mentalities bent on war are eager to twist AI technology for their own lethal purposes.
At first, I admit, I was a bit flattered to learn that online entrepreneurs are selling study guides for my new book. I thought of CliffsNotes from long ago, helping fellow students who were short on time or interest to grasp the basics of notable works. Curiosity quickly won. I pulled out my credit card, paid $9.99 plus tax for one of the offerings, and awaited its arrival in the mail.
The thin booklet got off to a reasonable enough start, explaining with its first sentence, “The U.S. media coverage that makes it easier to sell wars to the public, as well as the often-hidden cost of civilian casualties from errant U.S. attacks, are all harshly criticized by journalist Solomon.” That wasn’t a bad sum-up of my book.
But the study guide’s second sentence was not nearly as good: “He guarantees that when Russia designated Ukrainian communities during the new attack, the U.S. media was everyone available and jumping into action with compassionate, piercing revealing.” Rereading that sentence a few times didn’t improve it, and I began to worry.
The computer-programmed assaults on the English language escalated. And so, the “war on terror” became the “battle on dread.” The Costs of War project at Brown University became “the Expenses of War project at Earthy Colored College.”
To the extent that meaning could be grasped, the next pages seemed to include some praise: My book “constructs a convincing case that an excessive number of mysteries are being kept from people in general.” What’s more, “the creator presents a sharp and provocative outline of the outcomes of the media’s horrifying disappointments in spreading the word.”
But the study guide also included mild criticism amid the odd wording: “Solomon might have offered a fairly more profound examination of why American newscasting neglects to satisfy its beliefs in covering war and the justifications for why political pioneers could feel a sense of urgency to deal with misdirection while tending to people in general.”
The computer-programmed assaults on the English language escalated. And so, the “war on terror” became the “battle on dread.” A key source of meticulous research that I cited in my book, the Costs of War project at Brown University, became “the Expenses of War project at Earthy Colored College.”
At one point, my book’s actual title—“War Made Invisible”—shifted to “War Caused Imperceptible.” But the laughable malapropisms provided by artificial intelligence became more serious matters when I saw several dozen words forming badly mangled phrases—all attributed to me—inside quotation marks. I could imagine bleary-eyed students cramming on the night before a test or a term-paper deadline, reading the ostensible quotes and thinking that the author of my book must be an idiot.
Likewise, any would-be scholars seeking to glean the gist of the book’s themes in exchange for their $9.99 purchase will surely come away mystified at best after reading sentences like: “It’s totally unsuitable for writers to toe the conflict line for a really long time, and afterward, at last report, essentially, it tends to be informed years past the point of no return.”
I’m not among the authors who claim to never read reviews of their books. In fact, I remember them. So, I could recognize the uber-clumsy efforts of artificial intelligence that sifted through nearly a dozen reviews of War Made Invisible, lifting bits and pieces while weirdly substituting supposed synonyms to steer clear of plagiarism lawsuits.
So, let’s hear it for digital “free enterprise.” Or maybe that’s “unshackled business.” Nice AI work if you can get it.
Which brings us to a vastly more substantive matter. Artificial so-called intelligence is hardly immune to a dynamic that computer experts long ago dubbed “GIGO”—garbage in, garbage out. With AI, no matter how sophisticated it might seem, the consequences in war are apt to be horrific. Six decades after Martin Luther King Jr. warned of “guided missiles and misguided men,” the missiles are even more terrible, the people ordering launches are no less misguided, and the mentalities bent on war are eager to twist AI technology for their own lethal purposes.
A couple of weeks ago, the Department of Defense (DOD) announced “the establishment of a generative artificial intelligence task force, an initiative that reflects the DOD’s commitment to harnessing the power of artificial intelligence in a responsible and strategic manner.”
If they were still alive, the 4.5 million people who have died as direct and indirect results of U.S. wars since 9/11 might doubt how “responsible” the Defense Department’s manner has been.
Let’s hope that the people running the Pentagon’s task force for artificial intelligence didn’t graduate from Earthy Colored College.
Dear Common Dreams reader, It’s been nearly 30 years since I co-founded Common Dreams with my late wife, Lina Newhouser. We had the radical notion that journalism should serve the public good, not corporate profits. It was clear to us from the outset what it would take to build such a project. No paid advertisements. No corporate sponsors. No millionaire publisher telling us what to think or do. Many people said we wouldn't last a year, but we proved those doubters wrong. Together with a tremendous team of journalists and dedicated staff, we built an independent media outlet free from the constraints of profits and corporate control. Our mission has always been simple: To inform. To inspire. To ignite change for the common good. Building Common Dreams was not easy. Our survival was never guaranteed. When you take on the most powerful forces—Wall Street greed, fossil fuel industry destruction, Big Tech lobbyists, and uber-rich oligarchs who have spent billions upon billions rigging the economy and democracy in their favor—the only bulwark you have is supporters who believe in your work. But here’s the urgent message from me today. It's never been this bad out there. And it's never been this hard to keep us going. At the very moment Common Dreams is most needed, the threats we face are intensifying. We need your support now more than ever. We don't accept corporate advertising and never will. We don't have a paywall because we don't think people should be blocked from critical news based on their ability to pay. Everything we do is funded by the donations of readers like you. When everyone does the little they can afford, we are strong. But if that support retreats or dries up, so do we. Will you donate now to make sure Common Dreams not only survives but thrives? —Craig Brown, Co-founder |
Norman Solomon is the national director of RootsAction.org and executive director of the Institute for Public Accuracy. The paperback edition of his latest book, War Made Invisible: How America Hides the Human Toll of Its Military Machine, includes an afterword about the Gaza war.
At first, I admit, I was a bit flattered to learn that online entrepreneurs are selling study guides for my new book. I thought of CliffsNotes from long ago, helping fellow students who were short on time or interest to grasp the basics of notable works. Curiosity quickly won. I pulled out my credit card, paid $9.99 plus tax for one of the offerings, and awaited its arrival in the mail.
The thin booklet got off to a reasonable enough start, explaining with its first sentence, “The U.S. media coverage that makes it easier to sell wars to the public, as well as the often-hidden cost of civilian casualties from errant U.S. attacks, are all harshly criticized by journalist Solomon.” That wasn’t a bad sum-up of my book.
But the study guide’s second sentence was not nearly as good: “He guarantees that when Russia designated Ukrainian communities during the new attack, the U.S. media was everyone available and jumping into action with compassionate, piercing revealing.” Rereading that sentence a few times didn’t improve it, and I began to worry.
The computer-programmed assaults on the English language escalated. And so, the “war on terror” became the “battle on dread.” The Costs of War project at Brown University became “the Expenses of War project at Earthy Colored College.”
To the extent that meaning could be grasped, the next pages seemed to include some praise: My book “constructs a convincing case that an excessive number of mysteries are being kept from people in general.” What’s more, “the creator presents a sharp and provocative outline of the outcomes of the media’s horrifying disappointments in spreading the word.”
But the study guide also included mild criticism amid the odd wording: “Solomon might have offered a fairly more profound examination of why American newscasting neglects to satisfy its beliefs in covering war and the justifications for why political pioneers could feel a sense of urgency to deal with misdirection while tending to people in general.”
The computer-programmed assaults on the English language escalated. And so, the “war on terror” became the “battle on dread.” A key source of meticulous research that I cited in my book, the Costs of War project at Brown University, became “the Expenses of War project at Earthy Colored College.”
At one point, my book’s actual title—“War Made Invisible”—shifted to “War Caused Imperceptible.” But the laughable malapropisms provided by artificial intelligence became more serious matters when I saw several dozen words forming badly mangled phrases—all attributed to me—inside quotation marks. I could imagine bleary-eyed students cramming on the night before a test or a term-paper deadline, reading the ostensible quotes and thinking that the author of my book must be an idiot.
Likewise, any would-be scholars seeking to glean the gist of the book’s themes in exchange for their $9.99 purchase will surely come away mystified at best after reading sentences like: “It’s totally unsuitable for writers to toe the conflict line for a really long time, and afterward, at last report, essentially, it tends to be informed years past the point of no return.”
I’m not among the authors who claim to never read reviews of their books. In fact, I remember them. So, I could recognize the uber-clumsy efforts of artificial intelligence that sifted through nearly a dozen reviews of War Made Invisible, lifting bits and pieces while weirdly substituting supposed synonyms to steer clear of plagiarism lawsuits.
So, let’s hear it for digital “free enterprise.” Or maybe that’s “unshackled business.” Nice AI work if you can get it.
Which brings us to a vastly more substantive matter. Artificial so-called intelligence is hardly immune to a dynamic that computer experts long ago dubbed “GIGO”—garbage in, garbage out. With AI, no matter how sophisticated it might seem, the consequences in war are apt to be horrific. Six decades after Martin Luther King Jr. warned of “guided missiles and misguided men,” the missiles are even more terrible, the people ordering launches are no less misguided, and the mentalities bent on war are eager to twist AI technology for their own lethal purposes.
A couple of weeks ago, the Department of Defense (DOD) announced “the establishment of a generative artificial intelligence task force, an initiative that reflects the DOD’s commitment to harnessing the power of artificial intelligence in a responsible and strategic manner.”
If they were still alive, the 4.5 million people who have died as direct and indirect results of U.S. wars since 9/11 might doubt how “responsible” the Defense Department’s manner has been.
Let’s hope that the people running the Pentagon’s task force for artificial intelligence didn’t graduate from Earthy Colored College.
Norman Solomon is the national director of RootsAction.org and executive director of the Institute for Public Accuracy. The paperback edition of his latest book, War Made Invisible: How America Hides the Human Toll of Its Military Machine, includes an afterword about the Gaza war.
At first, I admit, I was a bit flattered to learn that online entrepreneurs are selling study guides for my new book. I thought of CliffsNotes from long ago, helping fellow students who were short on time or interest to grasp the basics of notable works. Curiosity quickly won. I pulled out my credit card, paid $9.99 plus tax for one of the offerings, and awaited its arrival in the mail.
The thin booklet got off to a reasonable enough start, explaining with its first sentence, “The U.S. media coverage that makes it easier to sell wars to the public, as well as the often-hidden cost of civilian casualties from errant U.S. attacks, are all harshly criticized by journalist Solomon.” That wasn’t a bad sum-up of my book.
But the study guide’s second sentence was not nearly as good: “He guarantees that when Russia designated Ukrainian communities during the new attack, the U.S. media was everyone available and jumping into action with compassionate, piercing revealing.” Rereading that sentence a few times didn’t improve it, and I began to worry.
The computer-programmed assaults on the English language escalated. And so, the “war on terror” became the “battle on dread.” The Costs of War project at Brown University became “the Expenses of War project at Earthy Colored College.”
To the extent that meaning could be grasped, the next pages seemed to include some praise: My book “constructs a convincing case that an excessive number of mysteries are being kept from people in general.” What’s more, “the creator presents a sharp and provocative outline of the outcomes of the media’s horrifying disappointments in spreading the word.”
But the study guide also included mild criticism amid the odd wording: “Solomon might have offered a fairly more profound examination of why American newscasting neglects to satisfy its beliefs in covering war and the justifications for why political pioneers could feel a sense of urgency to deal with misdirection while tending to people in general.”
The computer-programmed assaults on the English language escalated. And so, the “war on terror” became the “battle on dread.” A key source of meticulous research that I cited in my book, the Costs of War project at Brown University, became “the Expenses of War project at Earthy Colored College.”
At one point, my book’s actual title—“War Made Invisible”—shifted to “War Caused Imperceptible.” But the laughable malapropisms provided by artificial intelligence became more serious matters when I saw several dozen words forming badly mangled phrases—all attributed to me—inside quotation marks. I could imagine bleary-eyed students cramming on the night before a test or a term-paper deadline, reading the ostensible quotes and thinking that the author of my book must be an idiot.
Likewise, any would-be scholars seeking to glean the gist of the book’s themes in exchange for their $9.99 purchase will surely come away mystified at best after reading sentences like: “It’s totally unsuitable for writers to toe the conflict line for a really long time, and afterward, at last report, essentially, it tends to be informed years past the point of no return.”
I’m not among the authors who claim to never read reviews of their books. In fact, I remember them. So, I could recognize the uber-clumsy efforts of artificial intelligence that sifted through nearly a dozen reviews of War Made Invisible, lifting bits and pieces while weirdly substituting supposed synonyms to steer clear of plagiarism lawsuits.
So, let’s hear it for digital “free enterprise.” Or maybe that’s “unshackled business.” Nice AI work if you can get it.
Which brings us to a vastly more substantive matter. Artificial so-called intelligence is hardly immune to a dynamic that computer experts long ago dubbed “GIGO”—garbage in, garbage out. With AI, no matter how sophisticated it might seem, the consequences in war are apt to be horrific. Six decades after Martin Luther King Jr. warned of “guided missiles and misguided men,” the missiles are even more terrible, the people ordering launches are no less misguided, and the mentalities bent on war are eager to twist AI technology for their own lethal purposes.
A couple of weeks ago, the Department of Defense (DOD) announced “the establishment of a generative artificial intelligence task force, an initiative that reflects the DOD’s commitment to harnessing the power of artificial intelligence in a responsible and strategic manner.”
If they were still alive, the 4.5 million people who have died as direct and indirect results of U.S. wars since 9/11 might doubt how “responsible” the Defense Department’s manner has been.
Let’s hope that the people running the Pentagon’s task force for artificial intelligence didn’t graduate from Earthy Colored College.