List of All LLM Models

Discover and compare 500+ large language models with real-time rankings, benchmarks, and community votes.

AllenAI: Molmo2 8B

AllenAI: Molmo2 8B

By Ai2

Molmo2-8B is an open vision-language model developed by the Allen Institute for AI (Ai2) as part of the Molmo2 family, supporting image, video, and multi-image understanding and grounding. It is based on Qwen3-8B and uses SigLIP 2 as its vision backbone, outperforming other open-weight, open-data models on short videos, counting, and captioning, while remaining competitive on long-video tasks.

Release Date

09 Jan 2026

Context Size

36.86K

AllenAI: Olmo 3.1 32B Instruct

AllenAI: Olmo 3.1 32B Instruct

By Ai2

Olmo 3.1 32B Instruct is a large-scale, 32-billion-parameter instruction-tuned language model engineered for high-performance conversational AI, multi-turn dialogue, and practical instruction following. As part of the Olmo 3.1 family, this variant emphasizes responsiveness to complex user directions and robust chat interactions while retaining strong capabilities on reasoning and coding benchmarks. Developed by Ai2 under the Apache 2.0 license, Olmo 3.1 32B Instruct reflects the Olmo initiative’s commitment to openness and transparency.

Release Date

06 Jan 2026

Context Size

65.54K

ByteDance Seed: Seedream 4.5

ByteDance Seed: Seedream 4.5

By bytedance-seed

Seedream 4.5 is the latest in-house image generation model developed by ByteDance. Compared with Seedream 4.0, it delivers comprehensive improvements, especially in editing consistency, including better preservation of subject details, lighting, and color tone. It also enhances portrait refinement and small-text rendering. The model’s multi-image composition capabilities have been significantly strengthened, and both reasoning performance and visual aesthetics continue to advance, enabling more accurate and artistically expressive image generation. Pricing is $0.04 per output image, regardless of size.

Release Date

23 Dec 2025

Context Size

4.10K

ByteDance Seed: Seed 1.6 Flash

ByteDance Seed: Seed 1.6 Flash

By bytedance-seed

Seed 1.6 Flash is an ultra-fast multimodal deep thinking model by ByteDance Seed, supporting both text and visual understanding. It features a 256k context window and can generate outputs of up to 16k tokens.

Release Date

23 Dec 2025

Context Size

262.14K

ByteDance Seed: Seed 1.6

ByteDance Seed: Seed 1.6

By bytedance-seed

Seed 1.6 is a general-purpose model released by the ByteDance Seed team. It incorporates multimodal capabilities and adaptive deep thinking with a 256K context window.

Release Date

23 Dec 2025

Context Size

262.14K

MiniMax: MiniMax M2.1

MiniMax: MiniMax M2.1

By MiniMax

MiniMax-M2.1 is a lightweight, state-of-the-art large language model optimized for coding, agentic workflows, and modern application development. With only 10 billion activated parameters, it delivers a major jump in real-world capability while maintaining exceptional latency, scalability, and cost efficiency. Compared to its predecessor, M2.1 delivers cleaner, more concise outputs and faster perceived response times. It shows leading multilingual coding performance across major systems and application languages, achieving 49.4% on Multi-SWE-Bench and 72.5% on SWE-Bench Multilingual, and serves as a versatile agent “brain” for IDEs, coding tools, and general-purpose assistance. To avoid degrading this model's performance, MiniMax highly recommends preserving reasoning between turns. Learn more about using reasoning_details to pass back reasoning in our [docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#preserving-reasoning-blocks).

Release Date

23 Dec 2025

Context Size

196.61K

Z.ai: GLM 4.7

Z.ai: GLM 4.7

By Z.ai

GLM-4.7 is Z.ai’s latest flagship model, featuring upgrades in two key areas: enhanced programming capabilities and more stable multi-step reasoning/execution. It demonstrates significant improvements in executing complex agent tasks while delivering more natural conversational experiences and superior front-end aesthetics.

Release Date

22 Dec 2025

Context Size

202.75K

Google: Gemini 3 Flash Preview

Google: Gemini 3 Flash Preview

By Google

Gemini 3 Flash Preview is a high speed, high value thinking model designed for agentic workflows, multi turn chat, and coding assistance. It delivers near Pro level reasoning and tool use performance with substantially lower latency than larger Gemini variants, making it well suited for interactive development, long running agent loops, and collaborative coding tasks. Compared to Gemini 2.5 Flash, it provides broad quality improvements across reasoning, multimodal understanding, and reliability. The model supports a 1M token context window and multimodal inputs including text, images, audio, video, and PDFs, with text output. It includes configurable reasoning via thinking levels (minimal, low, medium, high), structured output, tool use, and automatic context caching. Gemini 3 Flash Preview is optimized for users who want strong reasoning and agentic behavior without the cost or latency of full scale frontier models.

Release Date

17 Dec 2025

Context Size

1.05M

Mistral: Mistral Small Creative

Mistral: Mistral Small Creative

By Mistral AI

Mistral Small Creative is an experimental small model designed for creative writing, narrative generation, roleplay and character-driven dialogue, general-purpose instruction following, and conversational agents.

Release Date

16 Dec 2025

Context Size

32.77K

AllenAI: Olmo 3.1 32B Think

AllenAI: Olmo 3.1 32B Think

By Ai2

Olmo 3.1 32B Think is a large-scale, 32-billion-parameter model designed for deep reasoning, complex multi-step logic, and advanced instruction following. Building on the Olmo 3 series, version 3.1 delivers refined reasoning behavior and stronger performance across demanding evaluations and nuanced conversational tasks. Developed by Ai2 under the Apache 2.0 license, Olmo 3.1 32B Think continues the Olmo initiative’s commitment to openness, providing full transparency across model weights, code, and training methodology.

Release Date

16 Dec 2025

Context Size

65.54K

Black Forest Labs: FLUX.2 Max

Black Forest Labs: FLUX.2 Max

By black-forest-labs

FLUX.2 [max] is the new top-tier image model from Black Forest Labs, pushing image quality, prompt understanding, and editing consistency to the highest level yet. Pricing is as follows, [per the docs](https://bfl.ai/pricing?category=flux.2): Input: We charge $0.03 for each megapixel on the input (i.e. reference images for editing) Output: The first generated megapixel is charged $0.07. Each subsequent megapixel is charged $0.03.

Release Date

16 Dec 2025

Context Size

46.86K

Xiaomi: MiMo-V2-Flash

Xiaomi: MiMo-V2-Flash

By Xiaomi

MiMo-V2-Flash is an open-source foundation language model developed by Xiaomi. It is a Mixture-of-Experts model with 309B total parameters and 15B active parameters, adopting hybrid attention architecture. MiMo-V2-Flash supports a hybrid-thinking toggle and a 256K context window, and excels at reasoning, coding, and agent scenarios. On SWE-bench Verified and SWE-bench Multilingual, MiMo-V2-Flash ranks as the top #1 open-source model globally, delivering performance comparable to Claude Sonnet 4.5 while costing only about 3.5% as much. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config).

Release Date

14 Dec 2025

Context Size

262.14K

NVIDIA: Nemotron 3 Nano 30B A3B (free)

NVIDIA: Nemotron 3 Nano 30B A3B (free)

By nvidia

NVIDIA Nemotron 3 Nano 30B A3B is a small language MoE model with highest compute efficiency and accuracy for developers to build specialized agentic AI systems. The model is fully...

Release Date

14 Dec 2025

Context Size

256K

NVIDIA: Nemotron 3 Nano 30B A3B

NVIDIA: Nemotron 3 Nano 30B A3B

By Nvidia

NVIDIA Nemotron 3 Nano 30B A3B is a small language MoE model with highest compute efficiency and accuracy for developers to build specialized agentic AI systems. The model is fully open with open-weights, datasets and recipes so developers can easily customize, optimize, and deploy the model on their infrastructure for maximum privacy and security.

Release Date

14 Dec 2025

Context Size

262.14K

OpenAI: GPT-5.2 Chat

OpenAI: GPT-5.2 Chat

By OpenAI

GPT-5.2 Chat (AKA Instant) is the fast, lightweight member of the 5.2 family, optimized for low-latency chat while retaining strong general intelligence. It uses adaptive reasoning to selectively “think” on harder queries, improving accuracy on math, coding, and multi-step tasks without slowing down typical conversations. The model is warmer and more conversational by default, with better instruction following and more stable short-form reasoning. GPT-5.2 Chat is designed for high-throughput, interactive workloads where responsiveness and consistency matter more than deep deliberation.

Release Date

10 Dec 2025

Context Size

128K

OpenAI: GPT-5.2 Pro

OpenAI: GPT-5.2 Pro

By OpenAI

GPT-5.2 Pro is OpenAI’s most advanced model, offering major improvements in agentic coding and long context performance over GPT-5 Pro. It is optimized for complex tasks that require step-by-step reasoning, instruction following, and accuracy in high-stakes use cases. It supports test-time routing features and advanced prompt understanding, including user-specified intent like "think hard about this." Improvements include reductions in hallucination, sycophancy, and better performance in coding, writing, and health-related tasks.

Release Date

10 Dec 2025

Context Size

400K

OpenAI: GPT-5.2

OpenAI: GPT-5.2

By OpenAI

GPT-5.2 is the latest frontier-grade model in the GPT-5 series, offering stronger agentic and long context perfomance compared to GPT-5.1. It uses adaptive reasoning to allocate computation dynamically, responding quickly to simple queries while spending more depth on complex tasks. Built for broad task coverage, GPT-5.2 delivers consistent gains across math, coding, sciende, and tool calling workloads, with more coherent long-form answers and improved tool-use reliability.

Release Date

10 Dec 2025

Context Size

400K

Mistral: Devstral 2 2512

Mistral: Devstral 2 2512

By Mistral AI

Devstral 2 is a state-of-the-art open-source model by Mistral AI specializing in agentic coding. It is a 123B-parameter dense transformer model supporting a 256K context window. Devstral 2 supports exploring codebases and orchestrating changes across multiple files while maintaining architecture-level context. It tracks framework dependencies, detects failures, and retries with corrections—solving challenges like bug fixing and modernizing legacy systems. The model can be fine-tuned to prioritize specific languages or optimize for large enterprise codebases. It is available under a modified MIT license.

Release Date

09 Dec 2025

Context Size

262.14K

Sourceful: Riverflow V2 Max Preview

Sourceful: Riverflow V2 Max Preview

By sourceful

Riverflow V2 Max Preview is the most powerful variant of Sourceful's Riverflow V2 preview lineup. This preview version exceeds the performance of Riverflow 1 Family and is Sourceful's first unified text-to-image and image-to-image model family. Pricing is $0.075 per output image, regardless of size. Sourceful imposes a 4.5MB request size limit, therefore it is highly recommended to pass image URLs instead of Base64 data.

Release Date

08 Dec 2025

Context Size

8.19K

Sourceful: Riverflow V2 Standard Preview

Sourceful: Riverflow V2 Standard Preview

By sourceful

Riverflow V2 Standard Preview is the standard variant of Sourceful's Riverflow V2 preview lineup. This preview version exceeds the performance of Riverflow 1 Family and is Sourceful's first unified text-to-image and image-to-image model family. Pricing is $0.035 per output image, regardless of size. Sourceful imposes a 4.5MB request size limit, therefore it is highly recommended to pass image URLs instead of Base64 data.

Release Date

08 Dec 2025

Context Size

8.19K

Sourceful: Riverflow V2 Fast Preview

Sourceful: Riverflow V2 Fast Preview

By sourceful

Riverflow V2 Fast Preview is the fastest variant of Sourceful's Riverflow V2 preview lineup. This preview version exceeds the performance of Riverflow 1 Family and is Sourceful's first unified text-to-image and image-to-image model family. Pricing is $0.03 per output image, regardless of size. Sourceful imposes a 4.5MB request size limit, therefore it is highly recommended to pass image URLs instead of Base64 data.

Release Date

08 Dec 2025

Context Size

8.19K

Relace: Relace Search

Relace: Relace Search

By relace

The relace-search model uses 4-12 `view_file` and `grep` tools in parallel to explore a codebase and return relevant files to the user request. In contrast to RAG, relace-search performs agentic multi-step reasoning to produce highly precise results 4x faster than any frontier model. It's designed to serve as a subagent that passes its findings to an "oracle" coding agent, who orchestrates/performs the rest of the coding task. To use relace-search you need to build an appropriate agent harness, and parse the response for relevant information to hand off to the oracle. Read more about it in the [Relace documentation](https://docs.relace.ai/docs/fast-agentic-search/agent).

Release Date

08 Dec 2025

Context Size

256K

Z.ai: GLM 4.6V

Z.ai: GLM 4.6V

By Z.ai

GLM-4.6V is a large multimodal model designed for high-fidelity visual understanding and long-context reasoning across images, documents, and mixed media. It supports up to 128K tokens, processes complex page layouts and charts directly as visual inputs, and integrates native multimodal function calling to connect perception with downstream tool execution. The model also enables interleaved image-text generation and UI reconstruction workflows, including screenshot-to-HTML synthesis and iterative visual editing.

Release Date

08 Dec 2025

Context Size

131.07K

Nex AGI: DeepSeek V3.1 Nex N1

Nex AGI: DeepSeek V3.1 Nex N1

By nex-agi

DeepSeek V3.1 Nex-N1 is the flagship release of the Nex-N1 series — a post-trained model designed to highlight agent autonomy, tool use, and real-world productivity. Nex-N1 demonstrates competitive performance across all evaluation scenarios, showing particularly strong results in practical coding and HTML generation tasks.

Release Date

08 Dec 2025

Context Size

131.07K

EssentialAI: Rnj 1 Instruct

EssentialAI: Rnj 1 Instruct

By essentialai

Rnj-1 is an 8B-parameter, dense, open-weight model family developed by Essential AI and trained from scratch with a focus on programming, math, and scientific reasoning. The model demonstrates strong performance across multiple programming languages, tool-use workflows, and agentic execution environments (e.g., mini-SWE-agent).

Release Date

07 Dec 2025

Context Size

32.77K

Body Builder (beta)

Body Builder (beta)

By OpenRouter

Transform your natural language requests into structured OpenRouter API request objects. Describe what you want to accomplish with AI models, and Body Builder will construct the appropriate API calls. Example: "count to 10 using gemini and opus." This is useful for creating multi-model requests, custom model routers, or programmatic generation of API calls from human descriptions. **BETA NOTICE**: Body Builder is in beta, and currently free. Pricing and functionality may change in the future.

Release Date

05 Dec 2025

Context Size

128K

OpenAI: GPT-5.1-Codex-Max

OpenAI: GPT-5.1-Codex-Max

By OpenAI

GPT-5.1-Codex-Max is OpenAI’s latest agentic coding model, designed for long-running, high-context software development tasks. It is based on an updated version of the 5.1 reasoning stack and trained on agentic workflows spanning software engineering, mathematics, and research. GPT-5.1-Codex-Max delivers faster performance, improved reasoning, and higher token efficiency across the development lifecycle.

Release Date

04 Dec 2025

Context Size

400K

Amazon: Nova 2 Lite

Amazon: Nova 2 Lite

By Amazon

Nova 2 Lite is a fast, cost-effective reasoning model for everyday workloads that can process text, images, and videos to generate text. Nova 2 Lite demonstrates standout capabilities in processing documents, extracting information from videos, generating code, providing accurate grounded answers, and automating multi-step agentic workflows.

Release Date

02 Dec 2025

Context Size

1M

Mistral: Ministral 3 14B 2512

Mistral: Ministral 3 14B 2512

By Mistral AI

The largest model in the Ministral 3 family, Ministral 3 14B offers frontier capabilities and performance comparable to its larger Mistral Small 3.2 24B counterpart. A powerful and efficient language model with vision capabilities.

Release Date

02 Dec 2025

Context Size

262.14K

Mistral: Ministral 3 8B 2512

Mistral: Ministral 3 8B 2512

By Mistral AI

A balanced model in the Ministral 3 family, Ministral 3 8B is a powerful, efficient tiny language model with vision capabilities.

Release Date

02 Dec 2025

Context Size

262.14K

Showing page 6 of 25 with 737 models total