List of All LLM Models

Discover and compare 500+ large language models with real-time rankings, benchmarks, and community votes.

Qwen: Qwen3.6 27B

Qwen: Qwen3.6 27B

By Qwen

Qwen3.6 27B is a dense 27-billion-parameter language model from the Qwen Team at Alibaba, released in April 2026. It features hybrid multimodal capabilities — accepting text, image, and video inputs — and supports a 262,144-token context window. The model is designed for agentic coding and reasoning tasks, with particular strength in repository-level code comprehension, front-end development workflows, and multi-step problem solving. It includes a built-in thinking mode for extended reasoning and preserves thinking context across conversation history. Qwen3.6 27B supports 201 languages and dialects and is released under the Apache 2.0 license.

Release Date

27 Apr 2026

Context Size

262.14K

OpenAI: GPT-5.5 Pro

OpenAI: GPT-5.5 Pro

By OpenAI

GPT-5.5 Pro is OpenAI’s high-capability model optimized for deep reasoning and accuracy on complex, high-stakes workloads. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs, and is designed for long-horizon problem solving, agentic coding, and precise execution across multi-step workflows.

Release Date

24 Apr 2026

Context Size

1.05M

OpenAI: GPT-5.5

OpenAI: GPT-5.5

By OpenAI

GPT-5.5 is OpenAI’s frontier model designed for complex professional workloads, building on GPT-5.4 with stronger reasoning, higher reliability, and improved token efficiency on hard tasks. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs, enabling large-scale reasoning, coding, and multimodal workflows within a single system.

Release Date

24 Apr 2026

Context Size

1.05M

DeepSeek: DeepSeek V4 Pro

DeepSeek: DeepSeek V4 Pro

By DeepSeek

DeepSeek V4 Pro is a large-scale Mixture-of-Experts model from DeepSeek with 1.6T total parameters and 49B activated parameters, supporting a 1M-token context window. It is designed for advanced reasoning, coding, and long-horizon agent workflows, with strong performance across knowledge, math, and software engineering benchmarks. Built on the same architecture as DeepSeek V4 Flash, it introduces a hybrid attention system for efficient long-context processing. Reasoning efforts `high` and `xhigh` are supported; `xhigh` maps to max reasoning. It is well suited for complex workloads such as full-codebase analysis, multi-step automation, and large-scale information synthesis, where both capability and efficiency are critical.

Release Date

24 Apr 2026

Context Size

1.05M

DeepSeek: DeepSeek V4 Flash

DeepSeek: DeepSeek V4 Flash

By DeepSeek

DeepSeek V4 Flash is an efficiency-optimized Mixture-of-Experts model from DeepSeek with 284B total parameters and 13B activated parameters, supporting a 1M-token context window. It is designed for fast inference and high-throughput workloads, while maintaining strong reasoning and coding performance. The model includes hybrid attention for efficient long-context processing. Reasoning efforts `high` and `xhigh` are supported; `xhigh` maps to max reasoning. It is well suited for applications such as coding assistants, chat systems, and agent workflows where responsiveness and cost efficiency are important.

Release Date

24 Apr 2026

Context Size

1.05M

Google: Gemini 3.1 Flash TTS Preview

Google: Gemini 3.1 Flash TTS Preview

By Google

Gemini 3.1 Flash TTS Preview is a text-to-speech model from Google, and a substantial generational step up from Gemini 2.5 Flash TTS. It takes text input and produces audio output across 70+ languages — nearly 3× the language coverage of its predecessor. The headline addition is a system of 200+ inline audio tags (e.g. `[whispers]`, `[laughs]`, `[excited]`) that let developers steer delivery, emotion, and pacing mid-sentence, alongside a "director's chair" workflow in Google AI Studio for defining per-character Audio Profiles and scene-level context. It supports up to two speakers with independent voice and style configuration per speaker, outputs PCM audio at 24 kHz / 16-bit mono, and automatically watermarks all output with SynthID. Context window is 32k tokens.

Release Date

24 Apr 2026

Context Size

8.19K

Google: Veo 3.1 Fast

Google: Veo 3.1 Fast

By Google

Google's mid-tier video generation model balancing speed and quality. Veo 3.1 Fast generates high-quality video from text or image prompts with native synchronized audio, offering faster turnaround than Veo 3.1 at lower cost. Supports first-frame and last-frame conditioning, multiple resolutions and aspect ratios, and SynthID watermarking.

Release Date

24 Apr 2026

Context Size

0

Zyphra: Zonos v0.1 Transformer

Zyphra: Zonos v0.1 Transformer

By zyphra

Zonos v0.1 Transformer is a text-to-speech model from Zyphra built on a pure transformer architecture. It offers the same American and British English voice coverage as the Hybrid variant, and is suited for deployments where a transformer-only inference stack is preferred.

Release Date

23 Apr 2026

Context Size

4.10K

Zyphra: Zonos v0.1 Hybrid

Zyphra: Zonos v0.1 Hybrid

By zyphra

Zonos v0.1 Hybrid is a text-to-speech model from Zyphra built on a hybrid architecture. It produces English speech output with coverage across American and British accents in male and female voices. It is suited for English-language voice applications requiring accent and gender variety.

Release Date

23 Apr 2026

Context Size

4.10K

Sesame: CSM 1B

Sesame: CSM 1B

By sesame

CSM 1B is a conversational speech model from Sesame. It accepts text input and produces English speech output, with voice options spanning conversational and read-speech styles. At 1B parameters, it is suited for dialogue-oriented applications such as voice assistants and interactive agents.

Release Date

23 Apr 2026

Context Size

4.10K

Canopy Labs: Orpheus 3B

Canopy Labs: Orpheus 3B

By canopylabs

Orpheus 3B is an English text-to-speech model from Canopy Labs, fine-tuned for natural prosody and expressive delivery. It offers 7 preset voices and is suited for narration, voice assistants, and interactive applications where naturalistic speech is a priority.

Release Date

23 Apr 2026

Context Size

4.10K

hexgrad: Kokoro 82M

hexgrad: Kokoro 82M

By hexgrad

Kokoro 82M is a lightweight, open-weight text-to-speech model from hexgrad. It converts text to speech across 8 languages (American and British English, Spanish, French, Hindi, Italian, Japanese, Portuguese, and Chinese) using 54 preset voices organized by language and gender. At 82M parameters, it is well-suited for multilingual TTS deployments where footprint and cost efficiency matter.

Release Date

23 Apr 2026

Context Size

4.10K

Google: Veo 3.1 Lite

Google: Veo 3.1 Lite

By Google

Google's most cost-effective video generation model, designed for high-volume applications and rapid iteration. Veo 3.1 Lite generates 720p and 1080p video from text or image prompts with native synchronized audio at less than 50% of the cost of Veo 3.1 Fast. Supports 4–8 second clips in landscape (16:9) and portrait (9:16) formats, with SynthID watermarking. Ideal for content platforms, short-form video creation, and automated media generation.

Release Date

23 Apr 2026

Context Size

0

inclusionAI: Ling-2.6-1T

inclusionAI: Ling-2.6-1T

By inclusionai

Ling-2.6-1T is an instant (instruct) model from inclusionAI and the company’s trillion-parameter flagship, designed for real-world agents that require fast execution and high efficiency at scale. It uses a “fast thinking” approach to reduce costs to roughly a quarter of comparable models while maintaining top-tier performance. The model achieves state-of-the-art results on benchmarks such as AIME26 and SWE-bench Verified, and is well suited for advanced coding, complex reasoning, and large-scale agent workflows where both capability and efficiency are critical.

Release Date

23 Apr 2026

Context Size

262.14K

Tencent: Hy3 preview

Tencent: Hy3 preview

By tencent

Hy3 preview is a high-efficiency Mixture-of-Experts model from Tencent designed for agentic workflows and production use. It supports configurable reasoning levels across disabled, low, and high modes, allowing it to balance speed and depth depending on the task, while delivering strong code generation and reliable performance across multi-step, real-world workflows.

Release Date

22 Apr 2026

Context Size

262.14K

Xiaomi: MiMo-V2.5-Pro

Xiaomi: MiMo-V2.5-Pro

By Xiaomi

MiMo-V2.5-Pro is Xiaomi’s flagship model, delivering strong performance in general agentic capabilities, complex software engineering, and long-horizon tasks, with top rankings on benchmarks such as ClawEval, GDPVal, and SWE-bench Pro. It can independently and autonomously complete professional tasks that would take human experts days or weeks, involving more than a thousand tool calls. Its context length of up to 1M makes it well suited for integration with a wide range of agent frameworks.

Release Date

22 Apr 2026

Context Size

1.05M

Xiaomi: MiMo-V2.5

Xiaomi: MiMo-V2.5

By Xiaomi

MiMo-V2.5 is a native omnimodal model by Xiaomi. It delivers Pro-level agentic performance at roughly half the inference cost, while surpassing MiMo-V2-Omni in multimodal perception across image and video understanding tasks. Its 1M context window supports complete documents, extended conversations, and complex task contexts in a single pass, making it ideal for integration with agent frameworks where strong reasoning, rich perception, and cost efficiency all matter.

Release Date

22 Apr 2026

Context Size

1.05M

OpenAI: GPT-5.4 Image 2

OpenAI: GPT-5.4 Image 2

By OpenAI

[GPT-5.4](https://openrouter.ai/openai/gpt-5.4) Image 2 combines OpenAI's GPT-5.4 model with state-of-the-art image generation capabilities from GPT Image 2. It enables rich multimodal workflows, allowing users to seamlessly move between reasoning, coding, and visual generation within the same interaction.

Release Date

21 Apr 2026

Context Size

272K

inclusionAI: Ling-2.6-flash

inclusionAI: Ling-2.6-flash

By inclusionai

Ling-2.6-flash is an instant (instruct) model from inclusionAI with 104B total parameters and 7.4B active parameters, designed for real-world agents that require fast responses, strong execution, and high token efficiency. It delivers performance comparable to state-of-the-art models at a similar scale while significantly reducing token usage across coding, document processing, and lightweight agent workflows.

Release Date

21 Apr 2026

Context Size

262.14K

Pareto Code Router

Pareto Code Router

By OpenRouter

The Pareto Router maintains a tiered shortlist of strong coding models, ranked by [Artificial Analysis](https://artificialanalysis.ai/) coding percentiles. Set min_coding_score between 0 and 1 on the [pareto-router plugin](https://openrouter.ai/docs/guides/routing/routers/pareto-router#the-min_coding_score-parameter) to control how strong a coder you need; higher scores select stronger (and typically more expensive) models. If you omit min_coding_score, the router defaults to the High tier. Today scores map into three bands — Low, Medium, and High. More granular bands will be added over time, so the precise score you pick now lets you opt into finer control as the shortlist grows. Selecting Nitro from the variant dropdown ranks the models in your tier by measured throughput and routes each request to the fastest one, so you trade some model variety for lower latency. Read the [Pareto Router docs](https://openrouter.ai/docs/guides/routing/routers/pareto-router) for the full selection logic, fallback behavior, and how to customize routing.

Release Date

21 Apr 2026

Context Size

200K

Baidu: Qianfan-OCR-Fast (free)

Baidu: Qianfan-OCR-Fast (free)

By baidu

Qianfan-OCR-Fast is a domain-specific multimodal large model purpose-built for OCR. By leveraging specialized OCR training data while preserving versatile multimodal intelligence, it provides a powerful performance upgrade over Qianfan-OCR.

Release Date

20 Apr 2026

Context Size

65.54K

Baidu: Qianfan-OCR-Fast (free)

Baidu: Qianfan-OCR-Fast (free)

By baidu

Qianfan-OCR-Fast is a domain-specific multimodal large model purpose-built for OCR. By leveraging specialized OCR training data while preserving versatile multimodal intelligence, it provides a powerful performance upgrade over Qianfan-OCR.

Release Date

20 Apr 2026

Context Size

65.54K

Kling: Video O1

Kling: Video O1

By kwaivgi

Kling Video O1 is a video generation model from Kuaishou. It supports text and image inputs with video output, enabling text-to-video and image-to-video workflows. It is suited for cinematic content production, with first-frame and last-frame control for precise scene composition. It generates 5 or 10 second clips in 16:9, 9:16, or 1:1 aspect ratios.

Release Date

20 Apr 2026

Context Size

0

MiniMax: Hailuo 2.3

MiniMax: Hailuo 2.3

By MiniMax

Hailuo 2.3 is a video generation model from MiniMax. It accepts text prompts and reference images as input and generates video output, supporting both text-to-video and image-to-video workflows. It is suited for creative content production, cinematic scene generation, and character animation, with a focus on realistic motion and expressive character rendering.

Release Date

20 Apr 2026

Context Size

0

MoonshotAI: Kimi K2.6

MoonshotAI: Kimi K2.6

By moonshotai

Kimi K2.6 is Moonshot AI's next-generation multimodal model, designed for long-horizon coding, coding-driven UI/UX generation, and multi-agent orchestration. It handles complex end-to-end coding tasks across Python, Rust, and Go, and can convert prompts and visual inputs into production-ready interfaces. Its agent swarm architecture scales to hundreds of parallel sub-agents for autonomous task decomposition - delivering documents, websites, and spreadsheets in a single run without human oversight.

Release Date

20 Apr 2026

Context Size

262.14K

Mistral: Voxtral Mini TTS

Mistral: Voxtral Mini TTS

By Mistral AI

Voxtral Mini TTS is Mistral's text-to-speech model featuring zero-shot voice cloning and multilingual support. It converts text input into natural-sounding audio output.

Release Date

19 Apr 2026

Context Size

4.10K

OpenAI: GPT-4o Mini TTS

OpenAI: GPT-4o Mini TTS

By OpenAI

GPT-4o Mini TTS is OpenAI's cost-efficient text-to-speech model. It converts text input into natural-sounding audio output, supporting a variety of voices and tones.

Release Date

18 Apr 2026

Context Size

4.10K

Google: Gemini Embedding 2 Preview

Google: Gemini Embedding 2 Preview

By Google

Gemini Embedding 2 Preview is Google's first multimodal embedding model. We currently support mapping text and images into a unified vector space for semantic search and retrieval-augmented generation (RAG). It supports input context up to 8,192 tokens and flexible output dimensions from 128 to 3,072 (recommended: 768, 1536, or 3,072). Designed for cross-modal similarity — you can embed a text query and retrieve the most relevant images, or vice versa — making it well-suited for multimodal search, recommendation, and document understanding pipelines.

Release Date

17 Apr 2026

Context Size

8.19K

Anthropic: Claude Opus 4.7

Anthropic: Claude Opus 4.7

By Anthropic

Opus 4.7 is the next generation of Anthropic's Opus family, built for long-running, asynchronous agents. Building on the coding and agentic strengths of Opus 4.6, it delivers stronger performance on complex, multi-step tasks and more reliable agentic execution across extended workflows. It is especially effective for asynchronous agent pipelines where tasks unfold over time - large codebases, multi-stage debugging, and end-to-end project orchestration. Beyond coding, Opus 4.7 brings improved knowledge work capabilities - from drafting documents and building presentations to analyzing data. It maintains coherence across very long outputs and extended sessions, making it a strong default for tasks that require persistence, judgment, and follow-through. For users upgrading from earlier Opus versions, see our [official migration guide here](https://openrouter.ai/docs/guides/evaluate-and-optimize/model-migrations/claude-4-7)

Release Date

16 Apr 2026

Context Size

1M

ByteDance: Seedance 2.0

ByteDance: Seedance 2.0

By bytedance

Seedance 2.0 is a video generation model from ByteDance. It supports text-to-video, image-to-video with first and last frame control, and multimodal reference-to-video. It is particularly strong at preserving character consistency, visual style, and camera movement from reference material. The number of tokens is given by (height of output video * width of output video * duration * 24) / 1024

Release Date

15 Apr 2026

Context Size

0

Showing page 2 of 25 with 737 models total