Discover AI models for every task
Showing 1-16 of 16 models
by Deepseek
DeepSeek V3.1 is an optimized variant of DeepSeek V3 with enhanced chat capabilities. Offers excellent cost-efficiency with 685B MoE architecture and improved response quality for conversational tasks.
Highly efficient DeepSeek flagship engineered for fast, capable reasoning and low-cost inference.
DeepSeek V3.2 is the latest iteration of the DeepSeek V3 series with significant performance improvements. Features enhanced reasoning, coding capabilities, and better instruction following across diverse tasks.
DeepSeek R1 0528 is an upgraded 685B parameter reasoning model with significantly enhanced depth of reasoning and inference capabilities. Achieves 87.5% on AIME 2025 (up from 70%), 91.4% on AIME 2024, 73.3% on LiveCodeBench, and 1930 Codeforces rating. Features system prompt support, averages 23K thinking tokens per question for deeper analysis, and reduced hallucination rate. Released under MIT license supporting commercial use and distillation. Performance approaching O3 and Gemini 2.5 Pro levels.
by Black Forest Labs
Black Forest Labs FLUX.1 [dev] is a cutting-edge 12 billion parameter rectified flow transformer for text-to-image generation. Second only to FLUX.1 [pro] with strong prompt following matching closed-source alternatives. Features guidance distillation for efficient inference, high-resolution generation (1024x1024), accurate text rendering, and detailed composition. Supports both text-to-image and image-to-image generation. Open weights enable scientific research and innovative workflows.
Black Forest Labs FLUX.1 [dev] with LoRA adapter support. This variant enables fine-tuned generation with custom trained LoRA weights for specialized styles, characters, or concepts. Based on the full 12B parameter FLUX.1 [dev] model with all its capabilities including high-resolution generation, accurate text rendering, and detailed composition. Perfect for custom workflows and specialized image generation tasks.
by OpenAI
OpenAI Whisper Large V3 is a state-of-the-art automatic speech recognition model with 1550M parameters supporting 99 languages. Achieves 10-20% WER reduction compared to V2, trained on 1M hours weakly labeled + 4M hours pseudo-labeled audio. Features 128 Mel frequency bins (increased from 80), improved robustness to accents and background noise, and new Cantonese language support. Supports speech transcription and speech-to-English translation with sentence and word-level timestamps. Optimized with torch.compile for 4.5x speedup. Ideal for accessibility tools, multilingual transcription, and enterprise ASR applications.
OpenAI Whisper Large V3 Turbo is an optimized variant of Whisper V3 with significantly faster inference while maintaining high accuracy across 99 languages. Features architectural optimizations for reduced latency including faster encoder-decoder inference and efficient attention mechanisms. Delivers near-V3 accuracy with 2-3x speed improvement, ideal for real-time transcription applications, live subtitling, and high-throughput ASR workloads. Supports full multilingual capabilities, timestamps, and speech translation to English. Perfect for production deployments requiring both quality and speed.
by ZAI
ZAI GLM 5.1 is a 744B parameter Mixture-of-Experts language model built with the GLM‑MoE DSA architecture. It excels at agentic engineering, achieving state-of-the-art performance on benchmarks such as HLE with tools (52.3), SWE‑Bench Pro (58.4) and AIME 2026 (95.3). The model supports extensive tool use and long‑horizon reasoning, with a large context window of up to 128K tokens. It is released under the MIT license.
Black Forest Labs FLUX.2 [dev] is the latest generation text-to-image model with significant improvements over FLUX.1. Features enhanced prompt following, superior image quality, and faster generation. Built on the proven rectified flow transformer architecture with optimizations for better detail, composition, and text rendering. Excellent for creative workflows, concept art, and high-quality image generation with both text-to-image and image-to-image capabilities.
by Meta
Meta Llama 3.1 8B Instruct is an efficient multilingual instruction-tuned model optimized for dialogue and assistant use cases. With 8 billion parameters and 128K context length, it provides strong performance across general tasks, code generation, and multilingual understanding. Supports function calling and tool use with Grouped-Query Attention architecture. Ideal for deployment scenarios requiring lower compute resources while maintaining quality across English and 7 additional languages including German, French, Spanish, and Hindi.
Black Forest Labs FLUX.1 [schnell] is the fastest variant of the FLUX.1 family, optimized for rapid text-to-image generation with fewer inference steps. Built on the same 12B parameter rectified flow transformer architecture as FLUX.1 [dev] but distilled for maximum speed. Generates high-quality 1024x1024 images in 1-4 steps compared to 20-50 steps for standard models. Ideal for real-time applications, interactive tools, and high-throughput image generation scenarios. Apache 2.0 licensed for unrestricted use including commercial applications.
by MiniMax
MiniMax M2.1 is a state-of-the-art MoE model with 230B total / 10B active parameters, optimized for agentic coding and complex multi-step workflows. Excels at multilingual programming, tool use, and long-horizon planning. Matches Claude Sonnet 4.5 on code benchmarks and exceeds it in multilingual scenarios. Features 196K context window with FP8 efficiency. Released under Modified-MIT license for commercial use.
Meta's flagship 405B parameter model representing the pinnacle of open-source AI. Exceptional reasoning and comprehensive knowledge for demanding applications.
by Mistral
Devstral 2 123B is Mistral AI's flagship agentic coding model, featuring 123B parameters optimized for software engineering tasks. Achieves 72.2% on SWE-bench Verified and 61.3% on SWE-bench Multilingual. Excels at codebase exploration, multi-file editing, and agentic workflows with tool use. Supports 200K context window with enhanced function calling and structured output. Designed for IDE integration via Mistral Vibe CLI. Released under modified MIT license for unrestricted commercial use.
by NVIDIA
Nemotron Nano 12B V2 is a unified reasoning and chat model with controllable inference via /think and /no_think directives. Features hybrid Mamba-2 + MLP layers + 6 Attention layers architecture with 128K context. Achieves 76.25% AIME25, 97.75% MATH500, 70.79% LiveCodeBench, 66.98% BFCL v3. Supports runtime thinking budget control for accuracy-latency tradeoffs. Pre-trained on ~20T tokens with cutoff September 2024. Optimized for NVIDIA GPUs (A10G, H100, Jetson AGX Thor) with efficient Mamba-2 SSM for long-context handling. Includes native function calling and tool integration.