LLMsVideo TranslationImage GenerationVideo Generation
AI News

Independent coverage of the latest AI tool updates, releases, and comparisons.

Categories

  • AI LLMs
  • AI Video Translation
  • AI Image Generation
  • AI Video Generation

Company

  • About
  • Contact

Resources

  • Sitemap
  • AI Glossary
  • Tool Comparisons
  • Facts / Grounding
  • llms.txt
  • XML Sitemap
© 2026 AI News. Independent editorial coverage. Not affiliated with any AI company.
AI LLMs

DeepSeek V4 Expected in April — 1 Trillion Parameters, Native Multimodal

DeepSeek's V4 model targets 1T parameters with only 37B active per token, a 1M context window, and native image/video generation. Leaked benchmarks claim Claude Opus-level performance.

SM

Sarah Mueller

Monday, March 16, 2026·2 min read

DeepSeek V4 is expected to launch in April 2026, alongside Tencent's new Hunyuan model, according to Chinese tech outlet Whale Lab reported by Dataconomy. The model has been anticipated since mid-February, with multiple projected release windows passing without a public launch.

What We Know About V4

DeepSeek V4 is a ~1 trillion parameter Mixture-of-Experts model with only ~37B active parameters per token — meaning it can match the performance of much larger models while running on significantly less compute. The architecture includes a 1M-token context window powered by Engram conditional memory, a technology published on January 13 that enables efficient retrieval from extremely long contexts.

The model targets native multimodal generation: text, image, and video from a single architecture.

The Benchmark Claims

Leaked benchmarks claim 90% HumanEval and 80%+ SWE-bench Verified — which would match Claude Opus 4.6. These numbers are unverified and should be treated with appropriate skepticism until independent testing confirms them.

The Geopolitical Context

DeepSeek V4 is being optimized for domestic Chinese AI chips through partnerships with Huawei and Cambricon. This directly responds to US export controls on advanced semiconductors and aligns with China's push for AI hardware independence.

Meanwhile, OpenAI, Anthropic, and Google are cooperating to prevent model distillation — the technique DeepSeek previously used to train competitive models from Western frontier model outputs.

Our Take

If DeepSeek V4 delivers anywhere near its leaked benchmarks at the efficiency its architecture suggests, it will be the most cost-effective frontier model available. The 37B active parameters make it dramatically cheaper to run than Western alternatives. But "leaked benchmarks" from an unreleased model deserve exactly as much credibility as that phrase implies. Wait for the release.

Tools Mentioned

DeepSeekHigh-performance open-source LLMs with efficient training
Free (open source); API from $0.14/1M tokens
Claude (Anthropic)Safe, helpful AI assistant with extended context and reasoning
$20/mo (Pro)
GPT (OpenAI)Industry-leading large language models powering ChatGPT
$20/mo (ChatGPT Plus)

More in AI LLMs

AI LLMs

Meta Launches Muse Spark — Its First Closed-Source Model Targets 'Personal Superintelligence'

Meta Superintelligence Labs unveils Muse Spark with dual modes, 58% on Humanity's Last Exam, and multimodal reasoning. Breaking with tradition, the model is not open-source.

Alex Chen·Apr 8, 2026
AI LLMs

OpenAI, Anthropic, and Google Unite to Combat AI Model Copying From China

The three biggest Western AI labs are sharing information through the Frontier Model Forum to prevent Chinese competitors from extracting their models' capabilities.

Sarah Mueller·Apr 7, 2026
← Back to all news