DeepSeek V4: A Trillion Parameters, Zero Nvidia Chips
DeepSeek V4 launched March 2026 with a trillion parameters, optimized for Chinese chips. It costs 1/20th of GPT-5 and matches frontier performance.

DeepSeek V4 was released in early March 2026 as the first trillion-parameter AI model built to run on Chinese chips, not Nvidia's. The DeepSeek V4 release marked a turning point in the AI chip war: it's open-source, and its API costs about $0.14 per million input tokens. GPT-5 charges $2.50 for the same thing. That's an 18x price gap for models that, on several benchmarks, perform within spitting distance of each other.
The US spent three years trying to stop China from building frontier AI without American chips. The DeepSeek V4 release proved it doesn't need them.
When Was DeepSeek V4 Released?
DeepSeek V4 launched in early March 2026, roughly three months after V3. The release happened without fanfare — no press conference, no hype cycle. The model just appeared on the company's API and Hugging Face with full documentation and weights.
The Numbers Nobody Expected
A year ago, DeepSeek and Alibaba's Qwen held a combined 1% of the global AI market. Today that number is 15%. In the same period, OpenAI's ChatGPT app market share dropped from 69.1% to 45.3%, according to Apptopia data reported by Fortune.
That's not a gradual shift. That's a cliff.
V4 has roughly a trillion total parameters, but here's the trick: only 32 billion activate per token. It uses a Mixture-of-Experts architecture — think of it like a building with 100 offices but only 3 open at any given time. You get the knowledge of the whole building without paying to heat every room.
The result? V4 is actually more efficient per query than its predecessor V3, despite being 50% larger overall. Active parameters dropped from 37 billion to 32 billion. Bigger brain, lighter footprint.
The Chip Story Changes Everything
Here's where it gets interesting. V4 was optimized for Huawei's Ascend 910B/C chips and Cambricon processors. DeepSeek withheld early access from Nvidia and AMD entirely.
Read that again. The world's most anticipated AI model launched without support for the world's most popular AI chips.
This isn't spite. It's strategy. US export controls ban Nvidia's best hardware from reaching China. So DeepSeek built V4 to run on what China actually has. According to Digital Applied, it's "the first trillion-parameter model optimized entirely outside the NVIDIA ecosystem."
For Chinese enterprise buyers, this creates an obvious choice. Why chase banned American chips when the best new model runs natively on hardware you can actually buy?
The export controls were supposed to keep China a generation behind in AI. Instead, they gave Chinese chip makers a trillion-parameter reference customer.
What V4 Actually Does
The spec sheet reads like science fiction from two years ago. A million-token context window — up from 128,000 in V3. That means you can feed it an entire codebase of 50-100 files, or a 600-page legal contract, in a single request. No chunking. No lost context.
It's natively multimodal. Text, images, and video processed in one model, not three stitched together. Internal benchmarks claim 90% on HumanEval (coding) and 83.7% on SWE-Bench Verified, which would make it the best coding model on Earth. For context: GPT-5.2 High scores 80% on SWE-Bench.
Those numbers haven't been independently verified. DeepSeek's own benchmarks, like every lab's own benchmarks, deserve skepticism. But even if the real numbers are 5-10% lower, V4 is competing with models that cost 18 times more to use.
One tech consultancy ran a real test: 50,000 financial documents processed daily. GPT-5 cost $4,200/month. DeepSeek V4 cost $210. Accuracy within 2 percentage points.
It's Not Just DeepSeek
Alibaba's Qwen 3.5 launched February 16 with 397 billion parameters, Apache 2.0 licensing, 201-language support, and vision capabilities that beat GPT-5.2 on math-vision benchmarks. Qwen has now crossed 700 million cumulative downloads on Hugging Face — the most downloaded AI model family in the world, according to Xinhua.
Two Chinese model families went from footnotes to 15% market share in 12 months. The combined downloads exceed anything Meta, Google, or OpenAI have produced in the open-source space.
Then Qwen's creator left. Lin Junyang, the technical lead who built Qwen from scratch to 700 million downloads, resigned on March 3 — 15 days after the Qwen 3.5 launch. The post-training lead left the same day. The coding lead had already gone to Meta in January.
Alibaba built the most downloaded AI model on Earth. Then the team that built it walked out the door.
The Price War Nobody Can Win
DeepSeek V3 was reportedly trained for $5.6 million. V4's training cost hasn't been disclosed, but the architectural improvements suggest it's still in the single-digit millions. Compare that to the estimated $100 million or more per training run for GPT-5 and Claude Opus.
The math is brutal for Western AI companies. When your competitor's training run costs 2-5% of yours and the output is comparable, your pricing power evaporates. OpenAI can't charge $2.50 per million tokens forever when DeepSeek charges $0.14 for similar quality.
This isn't theoretical. Companies are already switching. That CTO's reaction — "Why are we paying 20x for this?" — is happening in boardrooms everywhere.
What This Actually Means
Three things are now true that weren't true a year ago.
First, frontier AI no longer requires American chips. V4 proves it. The export control strategy needs rethinking because its main effect has been to accelerate Chinese hardware adoption, not prevent Chinese AI progress.
Second, the price of intelligence just collapsed. When open-source models match proprietary ones at 1/20th the cost, "AI" stops being something only well-funded companies can deploy. A startup in Lagos or Lahore can now access the same model quality as a Fortune 500 company in San Francisco.
Third, the talent matters more than the brand. Qwen's creator built the most downloaded model family in history, then quit. DeepSeek's entire V4 team is anonymous — the company doesn't even list researchers on its papers. The people building frontier AI aren't staying at one company long enough to collect a second paycheck.
The AI race isn't being won by the richest lab. It's being won by whoever figures out how to do more with less. Right now, that's China. And the gap between what they charge and what Silicon Valley charges is the number every tech executive should have tattooed on their forearm.
The Albis Perception Gap Index scored the AI regulation story — closely connected to this competitive shift — at 8.0, with US and EU outlets framing the same technology race through completely different lenses. American coverage emphasizes security threats. European coverage emphasizes regulatory frameworks. Chinese coverage celebrates technological sovereignty. Same models. Same benchmarks. Three different stories about what they mean.
Sources & Verification
Based on 5 sources from 4 regions
- Particula TechEurope
- FortuneNorth America
- Digital AppliedInternational
- Awesome AgentsInternational
- AIBase / XinhuaAsia-Pacific
Keep Reading
DeepSeek Trained on Banned Chips, Then Locked Out the US
DeepSeek V4 launches in April after training on prohibited Nvidia hardware, then blocks US chipmakers from optimization access while giving Huawei a head start.
DeepSeek Trained on Banned Nvidia Chips. Crime or Trap?
US says DeepSeek used smuggled Nvidia Blackwell chips in Inner Mongolia. China says the export controls are illegitimate. PGI 6.55 — the same chips, two incompatible stories.
DeepSeek Used Nvidia's Best Chips. Nvidia Can't Touch It.
Export controls created a world where the customer gets the product and the supplier doesn't. How America's chip restrictions just inverted tech dominance.
Explore Perspectives
Get this delivered free every morning
The daily briefing with perspectives from 7 regions — straight to your inbox.
Free · Daily · Unsubscribe anytime
🔒 We never share your email