Artificial stupidity and me: If AI can do this to my book, what will it do in an actual war?

At first, I admit, I was a bit flattered to learn that online entrepreneurs were selling study guides for my new book. I thought of CliffsNotes from long ago, helping fellow students who were short on time or interest to grasp the basics of notable works. Curiosity quickly won out. I pulled out my credit card, paid $9.99 plus tax for one of the offerings, and awaited its arrival in the mail. 

The thin booklet got off to a reasonable enough start, explaining with its first sentence, “The U.S. media coverage that makes it easier to sell wars to the public, as well as the often-hidden cost of civilian casualties from errant U.S. attacks, are all harshly criticized by journalist Solomon.” That wasn’t a bad summing-up of my book.

But the study guide’s second sentence was not nearly as good: “He guarantees that when Russia designated Ukrainian communities during the new attack, the U.S. media was everyone available and jumping into action with compassionate, piercing revealing.” Rereading that sentence a few times didn’t improve it. I began to worry.

To the extent that any meaning could be grasped from them, the following pages seemed to include some praise: My book “constructs a convincing case that an excessive number of mysteries are being kept from people in general.” What’s more, “the creator presents a sharp and provocative outline of the outcomes of the media’s horrifying disappointments in spreading the word.”

But the study guide also included mild criticism, albeit so oddly worded as to be almost incomprehensible: “Solomon might have offered a fairly more profound examination of why American newscasting neglects to satisfy its beliefs in covering war and the justifications for why political pioneers could feel a sense of urgency to deal with misdirection while tending to people in general.”

The computer-programmed assaults on the English language escalated. And so the “war on terror” became the “battle on dread.” A key source of meticulous research that I cite in my book, the Costs of War project at Brown University, became “the Expenses of War project at Earthy Colored College.”

At one point, my book’s actual title — “War Made Invisible” — somehow became “War Caused Imperceptible.” But the laughable malapropisms provided by artificial intelligence became more serious matters when I saw several dozen words forming badly mangled phrases — all attributed to me — inside quotation marks. I could imagine bleary-eyed students cramming on the night before a test or a term-paper deadline, reading the ostensible quotes and thinking that the author of my book must be an idiot.

Likewise, any would-be scholars seeking to glean the gist of the book’s themes in exchange for their $9.99 purchase will surely come away mystified, at best, after encountering sentences like this: “It’s totally unsuitable for writers to toe the conflict line for a really long time, and afterward, at last report, essentially, it tends to be informed years past the point of no return.”

I’m not among the authors who claim to never read reviews of their books. In fact, I remember them. So I could recognize the uber-clumsy efforts of a large-language-model AI that had sifted through nearly a dozen reviews of “War Made Invisible,” lifting bits and pieces while weirdly substituting supposed synonyms to steer clear of plagiarism lawsuits.

So let’s hear it for digital “free enterprise.” Or maybe that’s “unshackled business.” Nice AI work if you can get it.


Want a daily wrap-up of all the news and commentary Salon has to offer? Subscribe to our morning newsletter, Crash Course.


That brings us, however, to a vastly more substantive matter. Artificial so-called intelligence is hardly immune to a dynamic that computer experts long ago dubbed “GIGO” — garbage in, garbage out. With AI, no matter how sophisticated it might seem, the consequences in war are apt to be horrific. Six decades after Martin Luther King Jr. warned of “guided missiles and misguided men,” the missiles are even more terrible, the people ordering launches are no less misguided, and the mentalities bent on war are eager to twist AI technology for their own lethal purposes.

A couple of weeks ago, the Department of Defense announced “the establishment of a generative artificial intelligence task force, an initiative that reflects the DoD’s commitment to harnessing the power of artificial intelligence in a responsible and strategic manner.”

If they were still alive, the 4.5 million people who have died as direct and indirect results of U.S. wars since 9/11 might cast doubt on how “responsible” the Defense Department’s manner has been.

Let’s hope that the people running the Pentagon’s AI task force didn’t graduate from Earthy Colored College.

Read more

about the ascendancy of AI

Comments

Leave a Reply

Skip to toolbar