1 / 3
Swipe to compare

DeepSeek V3.2 is a large-scale Mixture-of-Experts language model that harmonizes high computational efficiency with frontier-level reasoning and agentic tool-use performance. It introduces DeepSeek Sparse Attention (DSA), a fine-grained mechanism that reduces attention complexity from quadratic to linear, significantly cutting training and inference costs in long-context scenarios. Through scalable reinforcement learning post-training, it achieves performance comparable to GPT-5, with gold-medal results on the 2025 International Mathematical Olympiad and Olympiad in Informatics. The model also features a large-scale agentic task synthesis pipeline that improves instruction following and tool use in complex interactive environments.

Author
DeepSeekDeepSeek
Release Date
2025-12-01
Knowledge Cutoff
2025-03
License
Open Model
I/O Format
Context Length
164K / 164K
API I/O (1M)
$0.252 / $0.378
How to Use
API Access
Output Speed
47 tok/s
Arena Overall
1424
Intelligence Index
41.7
Coding Index
36.7
Math Index
92.0
LiveBench
49.8
ForecastBench
GPQA Diamond
84.0%
HLE
22.2%
MMLU-Pro
86.2%
AIME 2025
92.0%
MATH-500
LB Reasoning
44.3
LB Math
64.0
LB Data Analysis
45.0
LiveCodeBench
86.2%
LB Coding
75.7
LB Agentic
46.7
TAU2
90.6%
TerminalBench
35.6%
SciCode
38.9%
IFBench
60.7%
AA-LCR
0.7
Hallucination (HHEM)
6.3%
Factual Consistency (HHEM)
93.7%
LB Language
64.2
LB Instruction Following
23.1