Nemotron 3 Nano 30B A3B
Specifications
- Input
- Output
- Context window
- 128K tokens
- Released
- Dec 2025
Performance
- Speed
- 1 t/s
- TTFT
- 345 ms
- Latency
- —
- Intelligence
- —
Pricing
- Input
- €0.07 per 1M tokens
- Output
- €0.30 per 1M tokens
About this model
NVIDIA Nemotron 3 Nano is a highly efficient hybrid Mamba-Transformer MoE model with 30B total / 3.5B active parameters. Features 128K context window extensible to 1M tokens. Excels at agentic AI, reasoning, and tool calling tasks. Trained on 25T tokens with state-of-the-art efficiency. Supports English, German, French, Spanish, Italian, and Japanese. Open weights with commercial license.
Technical specifications
- Capabilities
- Input modalities
- Output modalities
- Reasoning
- Hybrid Default on
Knowledge horizon
Knowledge cutoff Jun 2025
Released Dec 2025
Today
Training to release 6 mo Since release 5 mo
See also
Add Model to Comparison
Search for a model to add