Back to Turning Points
Breakthrough

The DeepSeek Shock

The moment the market realized frontier AI capability was no longer geographically exclusive.

February 18, 2026
6 min read

By late 2024, the prevailing assumption in Western AI was that frontier capability required frontier capital. OpenAI had raised $6.6B. Anthropic had secured billions from Amazon and Google. NVIDIA's market cap exceeded $3 trillion on projected GPU demand. The scaling hypothesis, that performance tracked compute spend, underpinned a trillion-dollar infrastructure buildout.

This assumption was never unchallenged. Chinese AI labs had been publishing competitive research for years. Alibaba's Qwen, Baidu's ERNIE, and Zhipu AI's GLM models were all tracking Western benchmarks at lower price points. The open-source ecosystem, driven by Meta's Llama releases, Mistral in France, and a growing community of independent researchers, had been steadily closing the gap between proprietary and open models.

What US export controls on advanced NVIDIA chips did was force Chinese labs to prioritize efficiency over scale. DeepSeek, a research lab spun out of High-Flyer quantitative hedge fund and founded by Liang Wenfeng with $700M in backing, was among the most methodical. Operating under hardware restrictions, the team developed architectural innovations, particularly Mixture-of-Experts routing, that reduced training costs by orders of magnitude.

On December 26, 2024, DeepSeek released V3: a 671-billion-parameter MoE model that matched GPT-4-class benchmarks. The reported training cost was approximately $5.5 million, a fraction of publicly estimated training budgets for comparable Western models.

Three weeks later, they released R1.

V3 uses 671 billion parameters in a Mixture-of-Experts configuration, activating only a subset for each query. The reported training cost of $5.5M directly challenged the industry assumption that frontier models required $100M+ in compute. The model was trained under US export controls that limited access to the latest NVIDIA hardware.

R1 demonstrated that open-weight models could match proprietary reasoning capabilities. The accompanying chatbot app made these capabilities freely available to consumers, setting up the collision with Western AI companies that would follow within days.

DeepSeek R1 forced a fundamental repricing of the AI infrastructure thesis. If frontier models could be trained for $5.5M instead of $500M, the projected demand for GPU clusters — and the companies supplying them — needed recalculation. NVIDIA alone lost more market value in one day than the entire market cap of most S&P 500 companies.

The app offered free access to R1 reasoning capabilities that OpenAI charged $20/month for. The viral adoption demonstrated both consumer demand for AI reasoning tools and the competitive threat of Chinese AI products distributed directly to American users.

Italy's data protection authority imposed an emergency ban citing data storage in China. Within days, multiple US federal agencies banned DeepSeek from government networks. The bans focused not on the model's capabilities but on data sovereignty — where user conversations were stored and who could access them.

The MIT license release made DeepSeek's reasoning capabilities fully open. Any company or developer could use, modify, and distribute the model without restriction. The benchmark improvements demonstrated continued rapid progress despite export controls on hardware.

OpenAI sent a memo to the US House Select Committee on China claiming accounts associated with DeepSeek employees used obfuscated third-party routers to circumvent access restrictions and extract capabilities through distillation. The accusation raised fundamental questions about whether proprietary AI capabilities can be protected once deployed as a service.

Aftermath

DeepSeek R1 launched on January 20, 2025. It was a reasoning model with chain-of-thought capabilities comparable to OpenAI's o1. The model was open-weight. The chatbot app was free. Performance on math and coding benchmarks was competitive with the best proprietary offerings.

On January 27, the financial markets reacted. NVIDIA lost over $590 billion in market value in a single session. The broader NASDAQ shed approximately $1 trillion as investors reassessed whether the AI infrastructure buildout had been mispriced. It was among the largest single-day market capitalization declines in US stock market history.

The same day, DeepSeek's app surpassed ChatGPT as the most downloaded on the US App Store. It was the first Chinese AI product to reach that position.

Government responses followed within days. Italy's data protection authority banned DeepSeek on January 30, citing GDPR violations related to data storage in China. By February 1, NASA, the US Navy, the Pentagon, and other federal agencies had restricted it from government devices. The concerns centered on data sovereignty, not model capability.

Over the following year, DeepSeek continued publishing updates under increasingly open licenses. V3 moved to MIT license in March. R1-0528 improved AIME 2025 math scores from 70% to 87.5% while halving hallucination rates. By December 2025, V3.2-Speciale was benchmarking alongside Gemini 3.0 Pro.

In February 2026, OpenAI sent a memo to the US House Select Committee on China alleging that DeepSeek employees had used obfuscated third-party routers to circumvent access restrictions and extract capabilities through model distillation.

Industry Impact

The DeepSeek moment did not create the trends it exposed. Chinese AI labs had been progressing steadily. Open-source models had been gaining ground. Efficiency research was accelerating globally. What DeepSeek did was make the convergence visible and force the market to price it in.

  • The efficiency question. The $5.5M training cost figure challenged the assumption that frontier capability required $100M+ compute budgets. Whether through MoE architectures, better data curation, or training optimizations, DeepSeek demonstrated that algorithmic efficiency could substitute for hardware scale, at least partially.

  • The export control paradox. US chip restrictions were designed to slow Chinese AI development. By constraining hardware access, they incentivized exactly the kind of efficiency research that made hardware less decisive. The policy achieved its stated goal (limiting chip access) while potentially accelerating the outcome it aimed to prevent.

  • Open-weight distribution as strategy. DeepSeek's MIT license releases gave any developer access to frontier-competitive reasoning. Combined with Llama, Mistral, and Qwen, the open model ecosystem reached a critical mass that made proprietary containment structurally harder.

  • The infrastructure repricing. The market reaction was not about one model. It was a reassessment of the demand curve for GPU clusters. If frontier-competitive models can be trained at 10x to 100x lower cost, the growth assumptions underpinning NVIDIA, cloud providers, and data center REITs needed recalculation.

  • The distillation question. OpenAI's accusation raised a structural problem the industry had deferred. Once a model is deployed as a service, can its capabilities be prevented from extraction? If not, the moat around proprietary AI narrows with every API call.