List of All LLM Models

Discover and compare 500+ large language models with real-time rankings, benchmarks, and community votes.

Qwen: QwQ 32B Preview

Qwen: QwQ 32B Preview

By Qwen

QwQ-32B-Preview is an experimental research model focused on AI reasoning capabilities developed by the Qwen Team. As a preview release, it demonstrates promising analytical abilities while having several important limitations: 1. **Language Mixing and Code-Switching**: The model may mix languages or switch between them unexpectedly, affecting response clarity. 2. **Recursive Reasoning Loops**: The model may enter circular reasoning patterns, leading to lengthy responses without a conclusive answer. 3. **Safety and Ethical Considerations**: The model requires enhanced safety measures to ensure reliable and secure performance, and users should exercise caution when deploying it. 4. **Performance and Benchmark Limitations**: The model excels in math and coding but has room for improvement in other areas, such as common sense reasoning and nuanced language understanding.

Release Date

28 Nov 2024

Context Size

32.77K

Google: Gemini Experimental 1121

Google: Gemini Experimental 1121

By Google

Experimental release (November 21st, 2024) of Gemini.

Release Date

21 Nov 2024

Context Size

40.96K

EVA Qwen2.5 72B

EVA Qwen2.5 72B

By EVA-UNIT-01

EVA Qwen2.5 72B is a roleplay and storywriting specialist model. It's a full-parameter finetune of Qwen2.5-72B on mixture of synthetic and natural data. It uses Celeste 70B 0.1 data mixture, greatly expanding it to improve versatility, creativity and "flavor" of the resulting model.

Release Date

21 Nov 2024

Context Size

32K

OpenAI: GPT-4o (2024-11-20)

OpenAI: GPT-4o (2024-11-20)

By OpenAI

The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses. GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.

Release Date

20 Nov 2024

Context Size

128K

Mistral Large 2411

Mistral Large 2411

By Mistral AI

Mistral Large 2 2411 is an update of [Mistral Large 2](/mistralai/mistral-large) released together with [Pixtral Large 2411](/mistralai/pixtral-large-2411) It provides a significant upgrade on the previous [Mistral Large 24.07](/mistralai/mistral-large-2407), with notable improvements in long context understanding, a new system prompt, and more accurate function calling.

Release Date

19 Nov 2024

Context Size

131.07K

Mistral Large 2407

Mistral Large 2407

By Mistral AI

This is Mistral AI's flagship model, Mistral Large 2 (version mistral-large-2407). It's a proprietary weights-available model and excels at reasoning, code, JSON, chat, and more. Read the launch announcement [here](https://mistral.ai/news/mistral-large-2407/). It supports dozens of languages including French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese, and Korean, along with 80+ coding languages including Python, Java, C, C++, JavaScript, and Bash. Its long context window allows precise information recall from large documents.

Release Date

19 Nov 2024

Context Size

131.07K

Mistral: Pixtral Large 2411

Mistral: Pixtral Large 2411

By Mistral AI

Pixtral Large is a 124B parameter, open-weight, multimodal model built on top of [Mistral Large 2](/mistralai/mistral-large-2411). The model is able to understand documents, charts and natural images. The model is available under the Mistral Research License (MRL) for research and educational use, and the Mistral Commercial License for experimentation, testing, and production for commercial purposes.

Release Date

19 Nov 2024

Context Size

131.07K

xAI: Grok Vision Beta

xAI: Grok Vision Beta

By xAI

Grok Vision Beta is xAI's experimental language model with vision capability.

Release Date

19 Nov 2024

Context Size

8.19K

Google: Gemini Experimental 1114

Google: Gemini Experimental 1114

By Google

Gemini 11-14 (2024) experimental model features "quality" improvements.

Release Date

15 Nov 2024

Context Size

40.96K

Infermatic: Mistral Nemo Inferor 12B

Infermatic: Mistral Nemo Inferor 12B

By Infermatic

Inferor 12B is a merge of top roleplay models, expert on immersive narratives and storytelling. This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [anthracite-org/magnum-v4-12b](https://openrouter.ai/anthracite-org/magnum-v4-72b) as a base.

Release Date

13 Nov 2024

Context Size

32K

Qwen2.5 Coder 32B Instruct

Qwen2.5 Coder 32B Instruct

By Qwen

Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). Qwen2.5-Coder brings the following improvements upon CodeQwen1.5: - Significantly improvements in **code generation**, **code reasoning** and **code fixing**. - A more comprehensive foundation for real-world applications such as **Code Agents**. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies. To read more about its evaluation results, check out [Qwen 2.5 Coder's blog](https://qwenlm.github.io/blog/qwen2.5-coder-family/).

Release Date

11 Nov 2024

Context Size

32.77K

SorcererLM 8x22B

SorcererLM 8x22B

By rAIfle

SorcererLM is an advanced RP and storytelling model, built as a Low-rank 16-bit LoRA fine-tuned on [WizardLM-2 8x22B](/microsoft/wizardlm-2-8x22b). - Advanced reasoning and emotional intelligence for engaging and immersive interactions - Vivid writing capabilities enriched with spatial and contextual awareness - Enhanced narrative depth, promoting creative and dynamic storytelling

Release Date

08 Nov 2024

Context Size

16K

EVA Qwen2.5 32B

EVA Qwen2.5 32B

By EVA-UNIT-01

EVA Qwen2.5 32B is a roleplaying/storywriting specialist model. It's a full-parameter finetune of Qwen2.5-32B on mixture of synthetic and natural data. It uses Celeste 70B 0.1 data mixture, greatly expanding it to improve versatility, creativity and "flavor" of the resulting model.

Release Date

08 Nov 2024

Context Size

32K

TheDrummer: UnslopNemo 12B

TheDrummer: UnslopNemo 12B

By Drummer

UnslopNemo v4.1 is the latest addition from the creator of Rocinante, designed for adventure writing and role-play scenarios.

Release Date

08 Nov 2024

Context Size

32.77K

Anthropic: Claude 3.5 Haiku (2024-10-22)

Anthropic: Claude 3.5 Haiku (2024-10-22)

By Anthropic

Claude 3.5 Haiku features enhancements across all skill sets including coding, tool use, and reasoning. As the fastest model in the Anthropic lineup, it offers rapid response times suitable for applications that require high interactivity and low latency, such as user-facing chatbots and on-the-fly code completions. It also excels in specialized tasks like data extraction and real-time content moderation, making it a versatile tool for a broad range of industries. It does not support image inputs. See the launch announcement and benchmark results [here](https://www.anthropic.com/news/3-5-models-and-computer-use)

Release Date

04 Nov 2024

Context Size

200K

Anthropic: Claude 3.5 Haiku

Anthropic: Claude 3.5 Haiku

By Anthropic

Claude 3.5 Haiku features offers enhanced capabilities in speed, coding accuracy, and tool use. Engineered to excel in real-time applications, it delivers quick response times that are essential for dynamic tasks such as chat interactions and immediate coding suggestions. This makes it highly suitable for environments that demand both speed and precision, such as software development, customer service bots, and data management systems. This model is currently pointing to [Claude 3.5 Haiku (2024-10-22)](/anthropic/claude-3-5-haiku-20241022).

Release Date

04 Nov 2024

Context Size

200K

NeverSleep: Lumimaid v0.2 70B

NeverSleep: Lumimaid v0.2 70B

By NeverSleep

Lumimaid v0.2 70B is a finetune of [Llama 3.1 70B](/meta-llama/llama-3.1-70b-instruct) with a "HUGE step up dataset wise" compared to Lumimaid v0.1. Sloppy chats output were purged. Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).

Release Date

22 Oct 2024

Context Size

131.07K

Anthropic: Claude 3.5 Sonnet

Anthropic: Claude 3.5 Sonnet

By Anthropic

New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at: - Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding - Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights - Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone - Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems) #multimodal

Release Date

22 Oct 2024

Context Size

200K

Magnum v4 72B

Magnum v4 72B

By anthracite-org

This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet(https://openrouter.ai/anthropic/claude-3.5-sonnet) and Opus(https://openrouter.ai/anthropic/claude-3-opus). The model is fine-tuned on top of [Qwen2.5 72B](https://openrouter.ai/qwen/qwen-2.5-72b-instruct).

Release Date

22 Oct 2024

Context Size

16.38K

xAI: Grok Beta

xAI: Grok Beta

By xAI

Grok Beta is xAI's experimental language model with state-of-the-art reasoning capabilities, best for complex and multi-step use cases. It is the successor of [Grok 2](https://x.ai/blog/grok-2) with enhanced context length.

Release Date

20 Oct 2024

Context Size

131.07K

Mistral: Ministral 3B

Mistral: Ministral 3B

By Mistral AI

Ministral 3B is a 3B parameter model optimized for on-device and edge computing. It excels in knowledge, commonsense reasoning, and function-calling, outperforming larger models like Mistral 7B on most benchmarks. Supporting up to 128k context length, it’s ideal for orchestrating agentic workflows and specialist tasks with efficient inference.

Release Date

17 Oct 2024

Context Size

128K

Mistral: Ministral 8B

Mistral: Ministral 8B

By Mistral AI

Ministral 8B is an 8B parameter model featuring a unique interleaved sliding-window attention pattern for faster, memory-efficient inference. Designed for edge use cases, it supports up to 128k context length and excels in knowledge and reasoning tasks. It outperforms peers in the sub-10B category, making it perfect for low-latency, privacy-first applications.

Release Date

17 Oct 2024

Context Size

128K

Qwen: Qwen2.5 7B Instruct

Qwen: Qwen2.5 7B Instruct

By Qwen

Qwen2.5 7B is the latest series of Qwen large language models. Qwen2.5 brings the following improvements upon Qwen2: - Significantly more knowledge and has greatly improved capabilities in coding and mathematics, thanks to our specialized expert models in these domains. - Significant improvements in instruction following, generating long texts (over 8K tokens), understanding structured data (e.g, tables), and generating structured outputs especially JSON. More resilient to the diversity of system prompts, enhancing role-play implementation and condition-setting for chatbots. - Long-context Support up to 128K tokens and can generate up to 8K tokens. - Multilingual support for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more. Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).

Release Date

16 Oct 2024

Context Size

32.77K

NVIDIA: Llama 3.1 Nemotron 70B Instruct

NVIDIA: Llama 3.1 Nemotron 70B Instruct

By Nvidia

NVIDIA's Llama 3.1 Nemotron 70B is a language model designed for generating precise and useful responses. Leveraging [Llama 3.1 70B](/models/meta-llama/llama-3.1-70b-instruct) architecture and Reinforcement Learning from Human Feedback (RLHF), it excels in automatic alignment benchmarks. This model is tailored for applications requiring high accuracy in helpfulness and response generation, suitable for diverse user queries across multiple domains. Usage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).

Release Date

15 Oct 2024

Context Size

131.07K

xAI: Grok 2

xAI: Grok 2

By xAI

Grok 2 is xAI's frontier language model with state-of-the-art reasoning capabilities, best for complex and multi-step use cases. To use a faster version, see [Grok 2 Mini](/x-ai/grok-2-mini). For more information, see the [launch announcement](https://x.ai/blog/grok-2).

Release Date

12 Oct 2024

Context Size

32.77K

xAI: Grok 2 mini

xAI: Grok 2 mini

By xAI

Grok 2 Mini is xAI's fast, lightweight language model that offers a balance between speed and answer quality. To use the stronger model, see [Grok Beta](/x-ai/grok-beta). For more information, see the [launch announcement](https://x.ai/blog/grok-2).

Release Date

12 Oct 2024

Context Size

32.77K

Inflection: Inflection 3 Pi

Inflection: Inflection 3 Pi

By Inflection

Inflection 3 Pi powers Inflection's [Pi](https://pi.ai) chatbot, including backstory, emotional intelligence, productivity, and safety. It has access to recent news, and excels in scenarios like customer support and roleplay. Pi has been trained to mirror your tone and style, if you use more emojis, so will Pi! Try experimenting with various prompts and conversation styles.

Release Date

11 Oct 2024

Context Size

8K

Inflection: Inflection 3 Productivity

Inflection: Inflection 3 Productivity

By Inflection

Inflection 3 Productivity is optimized for following instructions. It is better for tasks requiring JSON output or precise adherence to provided guidelines. It has access to recent news. For emotional intelligence similar to Pi, see [Inflect 3 Pi](/inflection/inflection-3-pi) See [Inflection's announcement](https://inflection.ai/blog/enterprise) for more details.

Release Date

11 Oct 2024

Context Size

8K

Google: Gemini 1.5 Flash 8B

Google: Gemini 1.5 Flash 8B

By Google

Gemini Flash 1.5 8B is optimized for speed and efficiency, offering enhanced performance in small prompt tasks like chat, transcription, and translation. With reduced latency, it is highly effective for real-time and large-scale operations. This model focuses on cost-effective solutions while maintaining high-quality results. [Click here to learn more about this model](https://developers.googleblog.com/en/gemini-15-flash-8b-is-now-generally-available-for-use/). Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.dev/terms).

Release Date

03 Oct 2024

Context Size

1M

TheDrummer: Rocinante 12B

TheDrummer: Rocinante 12B

By Drummer

Rocinante 12B is designed for engaging storytelling and rich prose. Early testers have reported: - Expanded vocabulary with unique and expressive word choices - Enhanced creativity for vivid narratives - Adventure-filled and captivating stories

Release Date

30 Sep 2024

Context Size

32.77K

Showing page 18 of 25 with 737 models total