Language Machinery

Richard Hughes Gibson in The Hedgehog Review:

Consider a few of the bolder claims made by experts. Two years ago, Blaise Agüera y Arcas, vice president of Google Research, had already declared the end of the animal kingdom’s monopoly on language on the strength of Google’s experiments with large language models. LLMs, he argued, “illustrate for the first time the way that language understanding and intelligence can be dissociated from all the embodied and emotional characteristics we share with each other and with many other animals.” In a similar vein, the Stanford University computer scientist Christopher Manning has argued that if “meaning” constitutes “understanding of the network of connections between linguistic form and other things,” be they “objects in the world or other linguistic forms,” then “there can be no doubt” that LLMs can “learn meanings.” Again, the point is that humans have company. The philosopher Tobias Rees (among many others) has gone further, arguing that LLMs constitute a “far-reaching, epoch-making philosophical event” on par with the shift from the premodern conception of language as a divine gift to the modern notion of language as a distinctly human trait, even our defining one. On Rees’s telling, engineers at OpenAI, Google, and Facebook have become the new Descartes and Locke, “[rendering] untenable the idea that only humans have language” and thereby undermining the modern paradigm those philosophers inaugurated. LLMs, for Rees at least, signal modernity’s end.

More here.