Command Palette
Search for a command to run

DeepSeek V4 Pro

by Deepseek

Specifications

Input
Output
Context window
1M tokens
Released
Apr 2026

Performance

Speed
3 t/s
TTFT
2.4s
Latency
Intelligence

Pricing

Input
€2.02
per 1M tokens
Output
€4.03
per 1M tokens

About this model

DeepSeek V4 Pro is a 1.6 T parameter Mixture-of-Experts (MoE) chat model from DeepSeek AI. It features a hybrid attention architecture with compressed sparse and heavily compressed attention, supporting a context window of one million tokens. The model achieves 90.1 % EM on MMLU, 76.8 % Pass@1 on HumanEval, and 51.5 % EM on LongBench‑V2, demonstrating strong language, coding, and long‑context capabilities. It is released under the MIT License.

Technical specifications

Capabilities
Input modalities
Output modalities
Reasoning
Hybrid Default on

Knowledge horizon

Released Apr 2026
Today
Since release 1 mo

See also

Add Model to Comparison
Search for a model to add
Command Palette
Search for a command to run