DeepSeek to Release V4 This Week: 1-Trillion-Parameter Multimodal Model Optimized for Huawei Chips

Read in other languages: 한국어日本語
LLM Mar 3, 2026 By Insights AI 1 min read 5 views Source

Chinese AI research lab DeepSeek is set to release DeepSeek V4 this week, per sources cited by TechNode and the Financial Times on March 2, 2026. The model has been repeatedly delayed since mid-February, with its release now timed to coincide with the start of China's annual Two Sessions parliamentary meetings on March 4.

V4 is built on a Mixture-of-Experts (MoE) architecture with approximately 32 billion active parameters from a total of 1 trillion parameters. It is a native multimodal model—trained from the ground up on text, images, video, and audio—with a context window of up to 1 million tokens. Leaked benchmarks suggest V4 would score around 90% on HumanEval and above 80% on SWE-Bench Verified, putting it in the same tier as Claude Opus 4.6 and GPT-5.3 Codex on coding tasks (unverified).

A notable strategic decision: DeepSeek deliberately excluded Nvidia and AMD from the pre-release optimization pipeline, building V4's inference stack entirely around Huawei Ascend and Cambricon chips. This positions V4 as a flagship model for Chinese domestic AI hardware, reflecting a broader trend of reducing dependence on US export-controlled semiconductors.

Three new architectural innovations are claimed: Manifold-Constrained Hyper-Connections (training stability at scale), Engram Conditional Memory (efficient retrieval across million-token contexts), and an enhanced Lightning Indexer for sparse attention.

Read more at TechNode.

Share:

Related Articles

Comments (0)

No comments yet. Be the first to comment!

Leave a Comment

© 2026 Insights. All rights reserved.