News How Robots Learned to Write So Well

How Robots Learned to Write So Well

LITERARY THEORY FOR ROBOTS: How Computers Learned to Write, by Dennis Yi Tenen


In “Literary Theory for Robots,” Dennis Yi Tenen’s playful new book on artificial intelligence and how computers learned to write, one of his most potent examples arrives in the form of a tiny mistake.

Tenen draws links between modern-day chatbots, pulp-fiction plot generators, old-fashioned dictionaries and medieval prophecy wheels. Both the utopians (the robots will save us!) and the doomsayers (the robots will destroy us!) have it wrong, he argues. There will always be an irreducibly human aspect to language and learning — a crucial core of meaning that emerges not just from syntax but from experience. Without it, you just get the chatter of parrots, who, “according to Descartes in his ‘Mediations,’ merely repeated without understanding,” Tenen writes.

But Descartes didn’t write “Mediations”; Tenen must have meant “Meditations” — the missing “t” will slip past any spell-checker program because both words are perfectly legitimate. (The book’s index lists the title correctly.) This minuscule typo doesn’t have any bearing on Tenen’s argument; if anything, it bolsters the case he wants to make. Machines are becoming stronger and smarter, but we still decide what is meaningful. A human wrote this book. And, despite the robots in the title, it is meant for other humans to read.

Tenen, now a professor of English and comparative literature at Columbia, used to be a software engineer at Microsoft. He puts his disparate skill sets to use in a book that is surprising, funny and resolutely unintimidating, even as he smuggles in big questions about art, intelligence, technology and the future of labor. I suspect that the book’s small size — it’s under 160 pages — is part of the point. People are not indefatigable machines, relentlessly ingesting enormous volumes on enormous subjects. Tenen has figured out how to present a web of complex ideas at human scale.

To that end, he tells stories, starting with the 14th-century Arab scholar Ibn Khaldun, who chronicled the use of the prophecy wheel, and ending with a chapter on the 20th-century Russian mathematician Andrey Markov, whose probability analysis of letter sequences in Pushkin’s “Eugene Onegin” constituted a fundamental building block of generative A.I. (Regular players of the game Wordle intuit such probabilities all the time.) Tenen writes knowledgeably about the technological roadblocks that stymied earlier models of computer learning, before “the brute force required to process most everything published in the English language” was so readily available. He urges us to be alert. He also urges us not to panic.

“Intelligence evolves on a spectrum, ranging from ‘partial assistance’ to ‘full automation’,” Tenen writes, offering the example of an automatic transmission in a car. Driving an automatic in the 1960s must have been mind-blowing for people used to manual transmissions. An automatic worked by automating key decisions, downshifting on hills and sending less power to the wheels in bad weather. It removed the option to stall or grind your gears. It was “artificially intelligent,” even if nobody used those words for it. American drivers now take its magic for granted. It has been demystified.

As for the current debates over A.I., this book tries to demystify those, too. Instead of talking about A.I. as if it has a mind of its own, Tenen talks about the collaborative work that went into building it. “We employ a cognitive-linguistic shortcut by condensing and ascribing agency to the technology itself,” he writes. “It’s easier to say, ‘The phone completes my messages’ instead of ‘The engineering team behind the autocompletion tool writing software based on the following dozen research papers completes my messages.’”

Our common metaphors for A.I. are therefore misleading. Tenen says we ought to be “suspicious of all metaphors ascribing familiar human cognitive aspects to artificial intelligence. The machine thinks, talks, explains, understands, writes, feels, etc., by analogy only.” This is why so much of his book revolves around questions of language. Language allows us to communicate and to understand one another. But it also allows for deception and misunderstanding. Tenen wants us to “unwind the metaphor” of A.I. — a proposal that might look like an English professor’s hobbyhorse on first glance but turns out to be entirely apt. A metaphor that is too general can make us complacent. Our sense of possibility is shaped by the metaphors we choose.

Text generators, whether in the form of 21st-century chatbots or 14th-century “letter magic,” have always faced the problem of “external validation,” Tenen writes. “Procedurally generated text can make grammatical sense, but might not always make sense sense.” Take Noam Chomsky’s famous example: “Colorless green ideas sleep furiously.” Anyone who has lived in the physical world would know that this syntactically flawless sentence is nonsense. Tenen keeps referring to the importance of “lived experience” because that describes our condition.

Tenen doesn’t deny that A.I. threatens much of what we call “knowledge work.” Nor does he deny that automating something also devalues it. But he also puts this another way: “Automation reduces barriers of entry, increasing the supply of goods for all.” Learning is cheaper now, and so having a big vocabulary or repertoire of memorized facts is no longer the competitive advantage it once was. “Today’s scribes and scholars can challenge themselves with more creative tasks,” he suggests. “Tasks that are tedious have been outsourced to the machines.”

I take his point, even if this prospect still seems bad to me, with an ever-shrinking sliver of the populace getting to do challenging, creative work while a once-flourishing ecosystem collapses. But Tenen also argues that we, as social beings, have agency, if only we allow ourselves to accept the responsibility that comes with it. “Individual A.I.s do pose real danger, given the ability to aggregate power in the pursuit of a goal,” he concedes. But the real danger comes “from our inability to hold technology makers responsible for their actions.” What if someone wanted to strap a jet engine to a car and see how it fared on the streets of a crowded city? Tenen says the answer is obvious: “Don’t do that.”

Why “Don’t do that” can seem easy in one realm but not another requires more thinking, more precision, more scrutiny — all qualities that fall by the wayside when we cower before A.I., treating the technology like a singular god instead of a multiplicity of machines built by a multiplicity of humans. Tenen leads by example, bringing his human intelligence to bear on artificial intelligence. By thinking through our collective habits of thought, he offers a meditation all his own.


LITERARY THEORY FOR ROBOTS: How Computers Learned to Write | By Dennis Yi Tenen | Norton | 158 pp. | $22

Exit mobile version