DeepSeek Breaks the AI Paradigm

Advisor Perspectives welcomes guest contributions. The views presented here do not necessarily represent those of Advisor Perspectives.

I’ve received emails from readers asking my thoughts on DeepSeek. I need to start with two warnings. First, the usual one: I’m a generalist value investor, not a technology specialist (last week I was analyzing a bank and an oil company), so my knowledge of AI models is superficial. Second, and more unusually, we don’t have all the facts yet.

But this story could represent a major step change in both AI and geopolitics. Here’s what we know:

DeepSeek—a year-old startup in China that spun out of a hedge fund—has built a fully functioning large language model (LLM) that performs on par with the latest AI models. This part of the story has been verified by the industry: DeepSeek has been tested and compared to other top LLMs. I’ve personally been playing with DeepSeek over the last few days, and the results it spit out were very similar to those produced by ChatGPT and Perplexity—only faster.

This alone is impressive, especially considering that just six months ago, Eric Schmidt (former Google CEO, and certainly no generalist) suggested China was two to three years behind the U.S. in AI.

But here’s the truly shocking—and unverified—part: DeepSeek claims they trained their model for only $5.6 million, while U.S. counterparts have reportedly spent hundreds of millions or even billions of dollars. That’s 20 to 200 times less.

The implications, if true, are stunning. Despite the U.S. government’s export controls on AI chips to China, DeepSeek allegedly trained its LLM on older-generation chips, using a small fraction of the computing power and electricity that its Western competitors have. While everyone assumed that AI’s future lay in faster, better chips—where the only real choice is Nvidia or Nvidia—this previously unknown company has achieved near parity with its American counterparts swimming in cash and datacenters full of the latest Nvidia chips. DeepSeek (allegedly) had huge compute constraints and thus had to use different logic, becoming more efficient with subpar hardware to achieve a similar result. In other words, this scrappy startup, in its quest to create a better AI “brain,” used brains where everyone else was focusing on brawn—it literally taught AI how to reason.