Chinese AI firm DeepSeek has emerged as a potential challenger to U.S. AI companies, demonstrating breakthrough models that claim to offer performance comparable to leading offerings at a fraction of the cost. The company’s mobile app, released in early January, has lately topped the App Store charts across major markets including the U.S., U.K., and China, but it hasn’t escaped doubts about whether its claims are true.
Founded in 2023 by Liang Wenfeng, the former chief of AI-driven quant hedge fund High-Flyer, DeepSeek’s models are open source and incorporate a reasoning feature that articulates its thinking before providing responses.
Wall Street’s reactions have been mixed. While brokerage firm Jefferies warns that DeepSeek’s efficient approach “punctures some of the capex euphoria” following recent spending commitments from Meta and Microsoft — each exceeding $60 billion this year — Citi is questioning whether such results were actually achieved without advanced GPUs.
Goldman Sachs sees broader implications, suggesting the development could reshape competition between established tech giants and startups by lowering barriers to entry.
Here’s how Wall Street analysts are reacting to DeepSeek, in their own words (emphasis ours):
Bank of America
Media chatter that DeepSeek’s R1 or R1-Zero models cost $5.6mn to development vs $1bn+ for alternative frontier models is misleading but also misses the bigger picture. DeepSeek noted the $5.6mn was the cost to train its previously released DeepSeek-V3 model using Nvidia H800 GPUs, but that the cost excluded other expenses related to research, experiments, architectures, algorithms and data. Our view is that more important than the significantly reduced cost and lower performance chips that DeepSeek used to develop its two newest models are the innovations introduced that enable more efficient (less costly) training and inference to occur in the first place. Lower AI compute costs should enable broader AI services from autos to smartphones.
Morgan Stanley
Bigger is no longer always smarter. DeepSeek demonstrates an alternative path to efficient model training than the current arm’s race among hyperscalers by significantly increasing the data quality and improving the model architecture. DeepSeek is now the lowest cost of LLM manufacturing, allowing frontier AI performance at a fraction of the cost with 9-13x lower price on output tokens vs. GPT-4o and Claude 3.5.
Why it matters. Frontier AI capabilities might be achievable without the massive computational resources previously thought necessary. Efficient resource use – with clever engineering and efficient training methods – could matter more than sheer computing power. This may inspire a wave of innovation in exploring cost-effective methods of AI development and deployment. This means that the ROI of LLM that is of today’s concern could improve meaningfully without giving away the quality or the time line for the deployment of AI applications. The achievement also suggests the democratization of AI by making sophisticated models more accessible to eventually drive greater adoption and proliferations of AI.
Bottom line. The restrictions on chips may end up acting as a meaningful tax on Chinese AI development but not a hard limit. China has demonstrated that cutting- edge AI capabilities can be achieved with significantly less hardware, defying conventional expectations of computing power requirements. A model that achieves frontier-grade results despite limited hardware access could mean a shift in the global AI landscape, redefining the competitive landscape of global AI enterprises, and fostering a new era of efficiency-driven progress.
Nomura
Although the first look on the DeepSeek’s effectiveness for training LLMs may lead to concerns for reduced hardware demand, we think large CSPs’ capex spending outlook would not change meaningfully in the near-term, as they need to stay in the competitive game, while they may accelerate the development schedule with the technology innovations. However, the market may become more anxious about the return on large AI investment, if there are no meaningful revenue streams in the near- term. Therefore, leading tech companies or CSPs may need to accelerate the AI adoptions and innovations; otherwise the sustainability of AI investment might be at risk. Another risk factor is the potential of more intensified competition between the US and China for AI leadership, which may lead to more technology restrictions and supply chain disruptions, in our view.
Jefferies
DeepSeek’s power implications for AI training punctures some of the capex euphoria which followed major commitments from Stargate and Meta last week. With DeepSeek delivering performance comparable to GPT-4o for a fraction of the computing power, there are potential negative implications for the builders, as pressure on AI players to justify ever increasing capex plans could ultimately lead to a lower trajectory for data center revenue and profit growth.
If smaller models can work well, it is potentially positive for smartphone. We are bearish on AI smartphone as AI has gained no traction with consumers. More hardware upgrade (adv pkg+fast DRAM) is needed to run bigger models on the phone, which will raise costs. AAPL’s model is in fact based on MoE, but 3bn data parameters are still too small to make the services useful to consumers. Hence DeepSeek’s success offers some hope but there is no impact on AI smartphone’s near-term outlook.
China is the only market that pursues LLM efficiency owing to chip constraint. Trump/Musk likely recognize the risk of further restrictions is to force China to innovate faster. Therefore, we think it likely Trump will relax the AI Diffusion policy.
Citi
While DeepSeek’s achievement could be groundbreaking, we question the notion that its feats were done without the use of advanced GPUs to fine tune it and/or build the underlying LLMs the final model is based on through the Distillation technique. While the dominance of the US companies on the most advanced AI models could be potentially challenged, that said, we estimate that in an inevitably more restrictive environment, US’ access to more advanced chips is an advantage. Thus, we don’t expect leading AI companies would move away from more advanced GPUs which provide more attractive $/TFLOPs at scale. We see the recent AI capex announcements like Stargate as a nod to the need for advanced chips.
Bernstein
In short, we believe that 1) DeepSeek DID NOT “build OpenAI for $5M”; 2) the models look fantastic but we don’t think they are miracles; and 3) the resulting Twitterverse panic over the weekend seems overblown.
Our own initial reaction does not include panic (far from it). If we acknowledge that DeepSeek may have reduced costs of achieving equivalent model performance by, say, 10x, we also note that current model cost trajectories are increasing by about that much every year anyway (the infamous “scaling laws…”) which can’t continue forever. In that context, we NEED innovations like this (MoE, distillation, mixed precision etc) if AI is to continue progressing. And for those looking for AI adoption, as semi analysts we are firm believers in the Jevons paradox (i.e. that efficiency gains generate a net increase in demand), and believe any new compute capacity unlocked is far more likely to get absorbed due to usage and demand increase vs impacting long term spending outlook at this point, as we do not believe compute needs are anywhere close to reaching their limit in AI. It also seems like a stretch to think the innovations being deployed by DeepSeek are completely unknown by the vast number of top tier AI researchers at the world’s other numerous AI labs (frankly we don’t know what the large closed labs have been using to develop and deploy their own models, but we just can’t believe that they have not considered or even perhaps used similar strategies themselves).
Goldman Sachs
With the latest developments, we also see 1) potential competition between capital-rich internet giants vs. start-ups, given lowering barriers to entry, especially with recent new models developed at a fraction of the cost of existing ones; 2) from training to more inferencing, with increased emphasis on post-training (including reasoning capabilities and reinforcement capabilities) that requires significantly lower computational resources vs. pre-training; and 3) the potential for further global expansion for Chinese players, given their performance and cost/price competitiveness.
We continue to expect the race for AI application/AI agents to continue in China, especially amongst To-C applications, where China companies have been pioneers in mobile applications in the internet era, e.g., Tencent’s creation of the Weixin (WeChat) super-app. Amongst To-C applications, ByteDance has been leading the way by launching 32 AI applications over the past year. Amongst them, Doubao has been the most popular AI Chatbot thus far in China with the highest MAU (c.70mn), which has recently been upgraded with its Doubao 1.5 Pro model. We believe incremental revenue streams (subscription, advertising) and eventual/sustainable path to monetization/positive unit economics amongst applications/agents will be key.
For the infrastructure layer, investor focus has centered around whether there will be a near-term mismatch between market expectations on AI capex and computing demand, in the event of significant improvements in cost/model computing efficiencies. For Chinese cloud/data center players, we continue to believe the focus for 2025 will center around chip availability and the ability of CSP (cloud service providers) to deliver improving revenue contribution from AI-driven cloud revenue growth, and beyond infrastructure/GPU renting, how AI workloads & AI related services could contribute to growth and margins going forward. We remain positive on long-term AI computing demand growth as a further lowering of computing/training/inference costs could drive higher AI adoption. See also Theme #5 of our key themes report for our base/bear scenarios for BBAT capex estimates depending on chip availability, where we expect aggregate capex growth of BBAT to continue in 2025E in our base case (GSe: +38% yoy) albeit at a slightly more moderate pace vs. a strong 2024 (GSe: +61% yoy), driven by ongoing investment into AI infrastructure.
J.P.Morgan
Above all, much is made of DeepSeek’s research papers, and of their models’ efficiency. It’s unclear to what extent DeepSeek is leveraging High-Flyer’s ~50k hopper GPUs (similar in size to the cluster on which OpenAI is believed to be training GPT-5), but what seems likely is that they’re dramatically reducing costs (inference costs for their V2 model, for example, are claimed to be 1/7 that of GPT-4 Turbo). Their subversive (though not new) claim – that started to hit the US AI names this week – is that “more investments do not equal more innovation.” Liang: “Right now I don’t see any new approaches, but big firms do not have a clear upper hand. Big firms have existing customers, but their cash-flow businesses are also their burden, and this makes them vulnerable to disruption at any time.” And when asked about the fact that GPT5 has still not been released: “OpenAI is not a god, they won’t necessarily always be at the forefront.”
UBS
Throughout 2024, the first year we saw massive AI training workload in China, more than 80-90% IDC demand was driven by AI training and concentrated in 1-2 hyperscaler customers, which translated to wholesale hyperscale IDC demand in relatively remote area (as power-consuming AI training is sensitive to utility cost rather than user latency).
If AI training and inference cost is significantly lower, we would expect more end users would leverage AI to improve their business or develop new use cases, especially retail customers. Such IDC demand means more focus on location (as user latency is more important than utility cost), and thus greater pricing power for IDC operators that have abundant resources in tier 1 and satellite cities. Meanwhile, a more diversified customer portfolio would also imply greater pricing power.
William Blair
From a semiconductor industry perspective, our initial take is that AI-focused semi companies are unlikely to see meaningful change to near-term demand trends given current supply constraints (around chips, memory, data center capacity, and power). Longer term, however, the continued pressure to lower the cost of compute—and the ability to reduce the cost of training and inference using new, more efficient algorithmic techniques—could result in lower capex than previously envisioned and lessen Nvidia’s dominance, especially if large-scale GPU clusters are not as critical to achieve frontier-level model performance as we thought. With still many unanswered questions and variables (what are the true costs of R1, what training data was used [only the model weights were open sourced], and how replicable are the results), we hesitate to come to any definitive conclusions regarding the future GenAI capex outlook (and whether DeepSeek has fundamentally altered it). That said, we recognize the hyper-sensitivity in the equity markets to overbuild risk, leading to today’s “shoot first and ask questions later” reaction.
We’ll update the story as more analysts react.
TechCrunch has an AI-focused newsletter! Sign up here to get it in your inbox every Wednesday.