DeepSeek V4 Preview Drops: China's AI Challenger Returns with Million-Word Memory

Image: Csis
Main Takeaway
DeepSeek's V4 preview offers million-word memory, but fresh Bloomberg tests show it still trails OpenAI's latest. MIT finds the model is leaner and proves Huawei chips can scale.
Jump to Key PointsSummary
What DeepSeek V4 actually delivers
DeepSeek dropped preview versions of its V4 model on Friday, exactly one year after its R1 reasoning model sent Nvidia stock tumbling and triggered a $1 trillion tech selloff. The Hangzhou startup says V4 can digest one million words of context while matching the performance of OpenAI's GPT-4, Google's Gemini, and Anthropic's Claude. TechCrunch argues the model "closes the gap" through architectural tweaks that make it more efficient than last year's V3.2.
But Bloomberg's first independent benchmarks tell a different story. After running V4 through standard reasoning, coding, and math evaluations, their testers found it still lags GPT-4.5 and Gemini 2.0 by a "clear margin." The gap is smaller than last year, yet the US lead hasn't vanished. DeepSeek's pricing playbook hasn't changed either: API access runs a fraction of OpenAI's rates, continuing the cost disruption that made the startup notorious.
How China built this without US chips
Here's where it gets spicy. DeepSeek tuned V4 specifically for Huawei's Ascend 910B processors, according to Fortune. That matters because US export controls have blocked Chinese companies from Nvidia's latest GPUs since 2022. The Verge notes this pokes holes in the assumption that Chinese AI shops need American hardware to compete.
MIT Technology Review adds a new wrinkle. Their teardown shows DeepSeek sanded down memory usage and bolted on custom low-precision math routines that let Ascend chips punch above their weight class. In plain English: they taught domestic silicon tricks normally reserved for Nvidia's top shelf. The result isn't just a lab demo—it's a signal that Huawei's foundry roadmap can keep scaling if engineers get creative enough.
CSIS analyst Gregory Allen points out the elephant in the room: if DeepSeek somehow obtained banned H100 chips, they'd be admitting to illegal activity. Instead, they've managed to squeeze near-frontier performance from domestic silicon. Reuters reports the model is "adapted to run on" Chinese chips, confirming the workarounds took serious engineering sweat.
What this means for the AI price war
DeepSeek's pricing strategy hasn't changed: undercut everyone. Fortune notes V4 continues the company's pattern of offering capabilities at "rock-bottom" rates. Bloomberg's testing confirms the model is "good enough" for most enterprise tasks, which may be all it needs to keep winning business on cost.
The bigger picture? Every dollar OpenAI and Google spend on newer, pricier chips looks a little less necessary if Huawei silicon can keep inching forward. MIT calls this "efficiency arbitrage"—when a cheaper, slightly weaker model wins by being dramatically less expensive to serve. That's the playbook DeepSeek keeps running, and so far, markets keep rewarding it.
Key Points
Bloomberg benchmarks show V4 still trails GPT-4.5 and Gemini 2.0 despite DeepSeek's parity claims
MIT teardown reveals custom optimizations that let Huawei Ascend chips handle million-word context efficiently
Pricing remains ultra-aggressive: API access costs a fraction of OpenAI's while being "good enough" for most business use
Huawei's foundry roadmap gains credibility as domestic chips scale without banned Nvidia hardware
Questions Answered
V4 handles 1 million words versus GPT-4's 128k limit and Claude's 200k maximum, representing roughly 8x improvement over current frontier models.
Yes, through DeepSeek's API and Hugging Face, though some enterprise features may have regional restrictions.
Specifically optimized for Huawei Ascend AI processors, avoiding banned Nvidia H100 chips entirely.
Roughly 75-80% cheaper per token, with API pricing around $10-15 per million tokens compared to OpenAI's $60+.
Benchmarks show near-parity on reasoning tasks, though full independent testing is still emerging.
Appears deliberate timing to mark the anniversary of their market-disrupting 2025 launch.
Source Reliability
45% of sources are highly trusted · Avg reliability: 71
Go deeper with Organic Intel
Simple AI systems for your life, work, and business. Each one includes copyable prompts, guides, and downloadable resources.
Explore Systems