by Ashutosh Jogalekar
In October last year, Charles Oppenheimer and I wrote a piece for Fast Company arguing that the only way to prevent an AI arms race is to open up the system. Drawing on a revolutionary early Cold War proposal for containing the spread of nuclear weapons, the Acheson-Lilienthal report, we argued that the foundational reason why security cannot be obtained through secrecy is because science and technology claim no real “secrets” that cannot be discovered if smart scientists and technologists are given enough time to find them. That was certainly the case with the atomic bomb. Even as American politicians and generals boasted that the United States would maintain nuclear supremacy for decades, perhaps forever, Russia responded with its first nuclear weapon merely four years after the end of World War II. Other countries like the United Kingdom, China and France soon followed. The myth of secrecy was shattered.
As if on cue after our article was written, in December 2024, a new large-language model (LLM) named DeepSeek v3 came out of China. DeepSeek v3 is a completely homegrown model built by a homegrown Chinese entrepreneur who was educated in China (that last point, while minor, is not unimportant: China’s best increasingly no longer are required to leave their homeland to excel). The model turned heads immediately because it was competitive with GPT-4 from OpenAI which many consider the state-of-the-art in pioneering LLM models. In fact, DeepSeek v3 is far beyond competitive in terms of critical parameters: GPT-4 used about 1 trillion training parameters, DeepSeek v3 used 671 billion; GPT-4 had 1 trillion tokens, DeepSeek v3 used almost 15 trillion. Most impressively, DeepSeek v3 cost only $5.58 million to train, while GPT-4 cost about $100 million. That’s a qualitatively significant difference: only the best-funded startups or large tech companies have $100 million to spend on training their AI model, but $5.58 million is well within the reach of many small startups.
Perhaps the biggest difference is that DeepSeek v3 is open-source while GPT-4 is not. The only other open source model from the United States is Llama, developed by Meta. If this feature of DeepSeek v3 is not ringing massive alarm bells in the heads of American technologists and political leaders, it should. It’s a reaffirmation of the central point that there are very few secrets in science and technology that cannot be discovered sooner or later by a technologically advanced country.
One might argue that DeepSeek v3 cost a fraction of the best LLM models to train because it stood on the shoulders of these giants, but that’s precisely the point: like other software, LLM models follow the standard rule of precipitously diminishing marginal cost. More importantly, the open-source, low-cost nature of DeepSeek v3 means that China now has the capability of capturing the world LLM market before the United States as millions of organizations and users make DeepSeek v3 the foundation on which to build their AI. Once again, the quest for security and technological primacy through secrecy would have proved ephemeral, just like it did for nuclear weapons.
What does the entry of DeepSeek v3 indicate in the grand scheme of things? It is important to dispel three myths and answer some key questions. Read more »















‘

Ryan Ruby is a novelist, translator, critic, and poet who lives, as I do, in Berlin. Back in the summer of 2018, I attended an event at TOP, an art space in Neukölln, where along with journalist Ben Mauk and translator Anne Posten, his colleagues at the Berlin Writers’ Workshop, he was reading from work in progress. Ryan read from a project he called Context Collapse, which, if I remember correctly, he described as a “poem containing the history of poetry.” But to my ears, it sounded more like an academic paper than a poem, with jargon imported from disciplines such as media theory, economics, and literary criticism. It even contained statistics, citations from previous scholarship, and explanatory footnotes, written in blank verse, which were printed out, shuffled up, and distributed to the audience. Throughout the reading, Ryan would hold up a number on a sheet of paper corresponding to the footnote in the text, and a voice from the audience would read it aloud, creating a spatialized, polyvocal sonic environment as well as, to be perfectly honest, a feeling of information overload. Later, I asked him to send me the excerpt, so I could delve deeper into what he had written at a slower pace than readings typically afford—and I’ve been looking forward to seeing the finished project ever since. And now that it is, I am publishing the first suite of excerpts from Context Collapse at Statorec, where I am editor-in-chief.![Henri Matisse created many paintings titled 'The Conversation'. This, from 2012, is of the artist with his wife, Amélie. [Hermitage Museum, St. Petersburg, Russia].](https://3quarksdaily.com/wp-content/uploads/2019/06/Matisse_Conversation-360x292.jpg)