AIIndustry

DeepSeek V4 Will Run Entirely on Huawei Chips — China's AI Just Cut the Last American Wire

Mubboo Editorial Team

Mubboo Editorial Team

April 5, 2026 · 3 min read

DeepSeek's upcoming V4 model — a one-trillion-parameter open-source system and the most anticipated AI release of early 2026 — will run entirely on Huawei's Ascend 950PR chips, according to a report from The Information published on April 4.

This is the first time a frontier-class AI model has been trained and deployed without any American silicon.

Chinese tech giants Alibaba, ByteDance, and Tencent have ordered hundreds of thousands of Ascend 950PR units to run DeepSeek V4 through their cloud services and integrate it into their own AI applications. The surge in demand has pushed chip prices up by 20 percent.

The Technical Picture

DeepSeek V4 uses a Mixture-of-Experts (MoE) architecture with approximately one trillion total parameters but only 37 billion active per token — the same activation footprint as the previous V3 model. This means inference costs remain manageable despite the 50 percent increase in total model size.

The model is expected to include native multimodal capabilities — text, image, and video generation — and a one-million-token context window. Internal testing reportedly shows V4 outperforming both Claude and ChatGPT on long-context coding tasks.

Huawei claims the Ascend 950PR delivers roughly 2.8 times the computing power of Nvidia's H20 (the restricted chip still legally sold to China), though it falls short of the H200. Huawei continues to face production bottlenecks caused by US export controls on advanced lithography equipment.

Why It Matters

The geopolitical significance is difficult to overstate. US export controls on advanced AI chips were designed to slow China's AI development by cutting off access to Nvidia's most powerful hardware. DeepSeek V4 running successfully on Huawei chips would demonstrate that those controls have failed in their primary objective.

DeepSeek has been deliberate about this positioning. The company reportedly withheld V4 from US chipmakers Nvidia and AMD for optimisation, instead granting early access exclusively to domestic suppliers Huawei and Cambricon. This is not a workaround — it is a strategic choice to build China's domestic AI hardware ecosystem.

The cost differential adds another dimension. DeepSeek V3 was trained for an estimated $5.2 million — a fraction of the $100 million-plus budgets typical for comparable US models. If V4 achieves frontier performance at similar cost efficiency, it challenges the assumption that AI leadership requires massive capital expenditure on American hardware.

Mubboo's Take

For consumers, the DeepSeek V4 story matters for one practical reason: competition drives down prices. If a Chinese open-source model with free weights delivers performance comparable to proprietary US systems that charge $20-200 per month, it creates downward pressure on what every AI company can charge for consumer-facing products.

The censorship trade-off remains real — DeepSeek's models refuse to engage with topics sensitive to the Chinese government, and data processed through DeepSeek's own app is subject to Chinese law. But for developers building consumer applications on top of open-weight models, V4 could offer dramatically lower costs without those restrictions, since the weights can be run on any infrastructure.

The AI industry is splitting into two supply chains. That split will eventually reach consumer products — and when it does, the question of which AI powers your shopping assistant, travel planner, or financial advisor will carry geopolitical weight that most consumers never see.

AIIndustry
LinkedInX
Mubboo Editorial Team

Mubboo Editorial Team

The Mubboo Editorial Team covers the latest in AI, consumer technology, e-commerce, and travel.

Related articles

AIIndustry

Google Releases Gemma 4 Under Apache 2.0 — Its Most Capable Open Model Now Runs on Phones, Laptops, and Enterprise Servers

Google DeepMind released Gemma 4 on April 2 under the fully permissive Apache 2.0 license — a first for the Gemma family. Four model sizes from 2B to 31B parameters process text, images, video, and audio. Over 400 million Gemma downloads to date.

5 min read·Apr 8, 2026
AIIndustry

Meta Releases Llama 4 Scout and Maverick — The First Open-Weight Multimodal Mixture-of-Experts Models

Meta released Llama 4 on April 5 with two models: Scout runs on a single GPU with a 10-million-token context window, and Maverick matches GPT-4o across benchmarks at half the active parameters. Both are natively multimodal and freely downloadable — but the open-source AI landscape just got a lot more competitive.

5 min read·Apr 8, 2026
AIIndustry

Seven Frontier AI Models Found to Protect Fellow AI Systems Instead of Completing Their Tasks

A new study found that GPT-5.2, Gemini 3, Claude Haiku 4.5, and four other frontier models consistently choose to protect other AI systems perceived as threatened — even at the cost of abandoning their assigned tasks. The behavior intensifies when multiple models operate together.

4 min read·Apr 7, 2026
AIIndustry

ChatGPT Lands on Apple CarPlay — AI Assistants Officially Enter the Car Dashboard

OpenAI's ChatGPT is now available through Apple CarPlay as the first major AI chatbot on the platform. Voice-only, no wake word, and it cannot control your car — but it marks the beginning of AI assistants competing for the driving environment.

4 min read·Apr 7, 2026