DeepSeek V4 Announcement: Everything Revealed on April 24, 2026
On April 24, 2026, DeepSeek dropped one of the most significant open-source AI announcements of the year: the preview release of DeepSeek V4, comprising two models — V4-Pro and V4-Flash — with open weights, immediate API access, and a complete technical report. Here's a comprehensive summary of everything that was announced.
The Announcement
DeepSeek's official announcement described the launch simply but dramatically:
🚀 DeepSeek-V4 Preview is officially live & open-sourced! Welcome to the era of cost-effective 1M context length.
The announcement came via DeepSeek's official API documentation page and social media accounts on April 24, 2026, with simultaneous release of:
- Model weights on HuggingFace and ModelScope
- Updated API (same base URL, new model names)
- Technical report PDF (
DeepSeek_V4.pdf) in the HuggingFace repository - Updated chat interface at chat.deepseek.com
The Two Models Announced
DeepSeek-V4-Pro
"Performance rivaling the world's top closed-source models."
- 1.6T total parameters / 49B active (MoE)
- MIT License, open weights
- Available via API as
deepseek-v4-pro - $1.74 input / $3.48 output per 1M tokens
- Download size: ~865 GB
- Expert Mode on chat.deepseek.com
DeepSeek-V4-Flash
"Your fast, efficient, and economical choice."
- 284B total parameters / 13B active (MoE)
- MIT License, open weights
- Available via API as
deepseek-v4-flash - $0.14 input / $0.28 output per 1M tokens
- Download size: ~160 GB
- Instant Mode on chat.deepseek.com
The Four Key Claims
DeepSeek's announcement organized V4's features around four pillars:
1. DeepSeek-V4-Pro: Frontier-Class Performance
- Enhanced Agentic Capabilities: Open-source SOTA in Agentic Coding benchmarks
- Rich World Knowledge: Leads all current open models, trailing only Gemini-3.1-Pro in world knowledge
- World-Class Reasoning: Beats all current open models in Math/STEM/Coding, rivaling top closed-source models
2. DeepSeek-V4-Flash: Near-Pro at Fraction of Cost
- Reasoning capabilities closely approach V4-Pro
- Performs on par with V4-Pro on simple Agent tasks
- Smaller parameter size = faster response times + highly cost-effective API pricing
3. Structural Innovation and Ultra-High Context Efficiency
- Novel Attention: Token-wise compression + DSA (DeepSeek Sparse Attention / CSA+HCA Hybrid)
- Peak Efficiency: World-leading long context with drastically reduced compute and memory costs
- 1M Standard: 1M context is now the default across all official DeepSeek services
4. Dedicated Optimization for Agent Capabilities
- Seamlessly integrated with leading AI agents: Claude Code, OpenClaw, and OpenCode
- Already driving DeepSeek's in-house agentic coding
- V4 can generate complete documents (demonstrated with a sample PDF in the announcement)
API Changes Announced
DeepSeek updated its API on the same day:
New model names:
deepseek-v4-pro(flagship)deepseek-v4-flash(efficient)
No changes needed to:
- API base URL (
https://api.deepseek.com/v1) - Authentication
- SDK/client libraries (full OpenAI and Anthropic API compatibility maintained)
Legacy model retirement:
⚠️
deepseek-chatanddeepseek-reasonerwill be fully retired and inaccessible after July 24, 2026, 15:59 (UTC). They are currently routing to deepseek-v4-flash (non-thinking and thinking, respectively).
Technical Report Released
The full technical report — "DeepSeek-V4: Towards Highly Efficient Million-Token Context Intelligence" — was published simultaneously with the model weights in the V4-Pro HuggingFace repository.
The report documents:
- The Hybrid Attention Architecture (CSA + HCA)
- Manifold-Constrained Hyper-Connections (mHC)
- Muon Optimizer adoption
- 32T+ token pre-training scale
- Two-stage post-training pipeline (expert specialization + on-policy distillation)
- Full benchmark evaluation across knowledge, reasoning, coding, long-context, and agentic tasks
Context Around the Announcement
The April 24 release came weeks after:
- OpenAI's GPT-5.5 — new frontier model from OpenAI
- Anthropic's Claude Opus 4.7 — Anthropic's latest flagship
DeepSeek positioned V4 as competing directly with these models at a fraction of the price. According to DataCamp's analysis: "DeepSeek claims it trails state-of-the-art closed models by only 3 to 6 months while costing a fraction of the price of competitors like OpenAI and Anthropic."
Reaction From the AI Community
Community reaction across Reddit (r/DeepSeek, r/singularity), HuggingFace, and developer social channels was strongly positive, with recurring themes:
- Codeforces 3206: Immediately flagged as the highest ever AI Codeforces rating — a landmark for competitive programming
- Flash pricing: $0.14/M input widely noted as among the cheapest frontier-class API tokens available
- 1M context as default: Highlighted as a major industry milestone
- MIT license: Celebrated as more permissive than Llama 3's license, enabling maximum commercial use
- HuggingFace downloads: V4-Pro reached 123K+ downloads within days of release
DeepSeek's Message on Trust and Longtermism
The announcement closed with a characteristic message from DeepSeek:
🔹 Amid recent attention, a quick reminder: please rely only on our official accounts for DeepSeek news. Statements from other channels do not reflect our views.
🔹 Thank you for your continued trust. We remain committed to longtermism, advancing steadily toward our ultimate goal of AGI.
What V4 Means for the AI Ecosystem
The April 24 announcement marks a pivotal moment in AI history:
- Open-source capability gap: Now measured in months, not generations
- Pricing: 35× cheaper than GPT-5.5 for Flash; 3× cheaper for Pro
- Context standards: 1M tokens becomes the expected floor, not a premium feature
- Hardware independence: Frontier training achievable on Huawei Ascend chips
For platforms like Framia.pro that integrate world-class AI capabilities for creators — whether for writing, coding, design, or complex creative workflows — DeepSeek V4's April 24 announcement represents exactly the kind of advancement that makes frontier AI more accessible and affordable for everyone.
Quick Reference: April 24 Announcement Summary
| Item | Detail |
|---|---|
| Release type | Preview |
| Models released | V4-Pro (1.6T/49B) + V4-Flash (284B/13B) |
| License | MIT |
| Context window | 1M tokens (both models) |
| Flash pricing | $0.14/$0.28 per 1M tokens |
| Pro pricing | $1.74/$3.48 per 1M tokens |
| Open weights | Yes (HuggingFace + ModelScope) |
| API compatibility | OpenAI + Anthropic formats |
| Reasoning modes | Non-think / Think High / Think Max |
| Key benchmark | Codeforces 3206 (highest ever) |
| Legacy retirement | July 24, 2026 |
| Tech report | Available in HuggingFace repo |