DeepSeek V4 Pro is DeepSeek's flagship open-source frontier model, released April 24, 2026, with 1.6 trillion total parameters (49B active) — the largest open-weight model to date. It introduces a hybrid attention architecture combining Compressed Sparse Attention (CSA) and Heavily Compressed Attention (HCA), reducing single-token FLOPs to 27% and KV cache to 10% compared to V3.2 at 1M-token context. Trained with mixed FP4/FP8 precision and Manifold-Constrained Hyper-Connections, it supports dual Thinking and Non-Thinking modes across a 1M-token context window. At launch, it scored 80.6% on SWE-bench Verified (tying Claude Opus 4.6), 93.5% on LiveCodeBench, and a 3206 Codeforces rating, while costing roughly one-seventh the price of comparable frontier models.
Reasoning|Proprietary Model
Knowledge Cutoff
2025-05
Input → Output Format
Context Memory
1.0MIN384KOUT
Arena Overall Score
1463
±9As of 2026-05-01
Overall Rank
No.24
4,176 Votes
Arena by Ability
Hard Prompts
1477±11No.35
Expert Knowledge
1476±28No.40
Instruction Following
1450±16No.30
Conversation Memory
1479±21No.21
Creative
1449±24No.16
Coding
1480±17No.57
Math
1444±32No.41
Arena by Occupation
Creative Writing
1448±19No.27
Social Sciences
1478±21No.31
Media
1434±22No.27
Business
1456±20No.31
Healthcare
1523±32🥇 No.1
Legal
1493±30No.10
Software
1479±14No.49
Mathematics
1452±32No.39
Source:Arena Intelligence
Overall
AA Intelligence Index
52%↑12%
LiveBench
74%↑14%
Reasoning & Math
GPQA Diamond
89%↑7%
HLE
36%↑18%
LB Reasoning
83%↑14%
LB Math
91%↑17%
LB Data
75%↑21%
Coding
AA Coding Index
48%↑11%
LB Coding
70%↓3%
LB Agentic
57%↑12%
TAU2
96%↑16%
TerminalBench
46%↑12%
SciCode
50%↑8%
Language & Instructions
IFBench
77%↑13%
AA-LCR
66%↑4%
LB Language
78%↑6%
LB IF
62%↑11%
Output Speed
Standard Mode
33tok/s↓44
First Output 1.16s