Mistral Models
Explore the Mistral language and embedding models available through our OpenAI Assistants API-compatible service.
Mistral: Mistral 7B Instruct
- Context Length:
- 32,768 tokens
- Architecture:
- text->text
- Max Output:
- 4,096 tokens
Pricing:
Introducing the Mistral 7B Instruct, a cutting-edge 7.3 billion parameter model designed to deliver exceptional performance and efficiency. Optimized for both speed and extended context handling, this model sets a new industry standard for advanced AI capabilities. As the latest iteration in the Mistral 7B Instruct series, it represents the forefront of innovation, offering enhanced precision and adaptability for a wide range of applications.
Mistral: Mistral 7B Instruct v0.3
- Context Length:
- 32,768 tokens
- Architecture:
- text->text
- Max Output:
- 4,096 tokens
Pricing:
Introducing a cutting-edge, high-performance 7.3 billion parameter model, meticulously optimized for speed and extended context length. This state-of-the-art model builds upon the foundation of Mistral 7B Instruct v0.2, delivering significant enhancements:
- Expanded Vocabulary: Now featuring an extended vocabulary of 32,768 tokens for improved language understanding and generation.
- v3 Tokenizer Support: Fully compatible with the latest v3 tokenizer for seamless integration and enhanced performance.
- Function Calling Capability: Equipped with support for function calling, enabling more dynamic and interactive applications. (Note: Function calling support may vary depending on the provider.)
Experience the next level of AI-driven solutions with this advanced model, designed to meet the demands of modern, high-stakes environments.
Mistral: Ministral 3B
- Context Length:
- 128,000 tokens
- Architecture:
- text->text
Pricing:
Ministral 3B is a cutting-edge 3-billion-parameter model specifically designed for on-device and edge computing applications. It delivers exceptional performance in areas such as knowledge retrieval, commonsense reasoning, and function-calling, often surpassing larger models like Mistral 7B across a wide range of benchmarks. With support for context lengths of up to 128k tokens, Ministral 3B is perfectly suited for managing complex agentic workflows and specialized tasks, offering highly efficient inference capabilities for real-world deployment.
Mistral: Ministral 8B
- Context Length:
- 128,000 tokens
- Architecture:
- text->text
Pricing:
The Ministral 8B is a cutting-edge 8-billion-parameter model engineered for exceptional performance and efficiency. Its innovative interleaved sliding-window attention mechanism enables faster inference and reduced memory usage, making it ideal for edge computing applications. With support for context lengths of up to 128k tokens, the model excels in knowledge-intensive and reasoning tasks, setting a new benchmark in the sub-10B parameter category. Ministral 8B outperforms comparable models, delivering superior results for low-latency, privacy-focused use cases where speed and efficiency are critical.
Mistral: Mistral Nemo
- Context Length:
- 131,072 tokens
- Architecture:
- text->text
Pricing:
Introducing a cutting-edge 12-billion-parameter model, developed by Mistral in collaboration with NVIDIA, featuring an impressive 128,000-token context length. This advanced model is designed to handle a wide range of languages, including English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi, making it a truly multilingual solution. Additionally, it supports function calling, enhancing its versatility for various applications. Released under the permissive Apache 2.0 license, this model offers both flexibility and accessibility for developers and researchers alike.
Mistral: Pixtral 12B
- Context Length:
- 4,096 tokens
- Architecture:
- text+image->text
Pricing:
Introducing Mistral AI's groundbreaking multi-modal model, capable of processing both text and images to generate text outputs. This innovative model represents a significant leap in AI capabilities, blending visual and textual understanding. Notably, Mistral AI has released the model's weights via torrent, making them accessible to the broader community. For more details, visit their official announcement: https://x.com/mistralai/status/1833758285167722836. This release marks a pivotal moment in AI development, showcasing Mistral AI's commitment to advancing multi-modal AI technologies.
Mistral: Mistral 7B Instruct v0.1
- Context Length:
- 4,096 tokens
- Architecture:
- text->text
Pricing:
Introducing a cutting-edge 7.3B parameter model that surpasses the Llama 2 13B across all benchmark metrics. This advanced model is meticulously optimized for enhanced processing speed and extended context length, delivering unparalleled efficiency and performance. Designed for both speed and scalability, it redefines the standards for high-performance AI models.
Mistral Small
- Context Length:
- 32,000 tokens
- Architecture:
- text->text
Pricing:
Boasting 22 billion parameters, Mistral Small v24.09 strikes an ideal balance between the compact efficiency of Mistral NeMo 12B and the advanced power of Mistral Large 2. This versatile model delivers a cost-effective, high-performance solution designed for seamless deployment across diverse platforms and environments. Mistral Small v24.09 excels in enhanced reasoning, expanded capabilities, and robust code generation and analysis. Additionally, it offers multilingual support, seamlessly handling English, French, German, Italian, and Spanish for a truly global reach.
Mistral: Mixtral 8x7B Instruct
- Context Length:
- 32,768 tokens
- Architecture:
- text->text
- Max Output:
- 4,096 tokens
Pricing:
Introducing Mixtral 8x7B Instruct, an advanced pretrained generative model developed by Mistral AI, specifically designed for chat and instruction-based applications. This cutting-edge Sparse Mixture of Experts (MoE) model integrates 8 specialized feed-forward networks, collectively comprising 47 billion parameters. Fine-tuned by Mistral AI, Mixtral 8x7B Instruct delivers exceptional performance and versatility for a wide range of conversational and instructional tasks.
Mistral Tiny
- Context Length:
- 32,000 tokens
- Architecture:
- text->text
Pricing:
This model is built on the Mistral-7B-v0.2 architecture and features enhanced fine-tuning, drawing inspiration from community-driven advancements to deliver improved performance over its predecessor, Mistral 7B. It is particularly well-suited for high-volume batch processing tasks where cost efficiency is a priority, and advanced reasoning capabilities are not a primary requirement.
Mistral: Codestral Mamba
- Context Length:
- 256,000 tokens
- Architecture:
- text->text
Pricing:
Introducing a cutting-edge 7.3 billion parameter Mamba-based model, expertly crafted for code generation and reasoning tasks. This advanced model delivers linear-time inference, enabling seamless processing of theoretically infinite sequence lengths, and boasts an expansive 256k token context window for handling complex tasks with ease. Optimized for speed, it ensures rapid responses, making it an ideal tool for boosting coding productivity.
Performance-wise, it rivals state-of-the-art transformer models in both code-related and reasoning tasks, offering exceptional accuracy and efficiency. Best of all, it’s available under the Apache 2.0 license, granting users the freedom to use, modify, and distribute it without restrictions. Whether you're a developer, researcher, or enthusiast, this model is designed to elevate your workflow and innovation.
Mistral: Mixtral 8x7B Instruct (nitro)
- Context Length:
- 32,768 tokens
- Architecture:
- text->text
Pricing:
Introducing Mixtral 8x7B Instruct, an advanced pretrained generative model developed by Mistral AI, specifically designed for chat and instruction-based applications. This cutting-edge Sparse Mixture of Experts (MoE) model integrates 8 specialized feed-forward networks, collectively comprising 47 billion parameters. Fine-tuned by Mistral AI, Mixtral 8x7B Instruct delivers exceptional performance and versatility, making it an ideal solution for sophisticated conversational and instructional tasks.
Mistral: Mixtral 8x22B Instruct
- Context Length:
- 65,536 tokens
- Architecture:
- text->text
Pricing:
Introducing Mistral's official instruct fine-tuned version of Mixtral 8x22B, a cutting-edge model designed to deliver exceptional performance with remarkable cost efficiency. With 39 billion active parameters out of a total 141 billion, this model sets a new standard for balancing power and affordability. Key features include:
- Advanced capabilities in mathematics, coding, and logical reasoning.
- A massive context length of 64k tokens, enabling deep and comprehensive understanding.
- Multilingual fluency in English, French, Italian, German, and Spanish.
For detailed performance benchmarks, explore the official launch announcement here.
Mistral Large
- Context Length:
- 128,000 tokens
- Architecture:
- text->text
Pricing:
Introducing Mistral Large 2 (version mistral-large-2407
), Mistral AI's flagship model—a proprietary, weights-available solution designed to excel in reasoning, coding, JSON handling, chat interactions, and more. Discover the full details in the official launch announcement here.
This versatile model supports a wide array of languages, including French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese, and Korean, as well as over 80 programming languages such as Python, Java, C, C++, JavaScript, and Bash. With its expansive context window, Mistral Large 2 ensures precise information retrieval from even the most extensive documents, making it a powerful tool for complex tasks.
Mistral Large 2407
- Context Length:
- 128,000 tokens
- Architecture:
- text->text
Pricing:
Introducing Mistral Large 2 (version mistral-large-2407), Mistral AI's flagship model. This proprietary, weights-available model is designed to excel in reasoning, coding, JSON handling, chat applications, and more. Discover the full details in the official launch announcement here.
Mistral Large 2 supports a wide array of languages, including French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese, and Korean. Additionally, it is proficient in over 80 programming languages such as Python, Java, C, C++, JavaScript, and Bash. With its expansive context window, the model ensures accurate information retrieval from extensive documents, making it a powerful tool for diverse applications.
Mistral Large 2411
- Context Length:
- 128,000 tokens
- Architecture:
- text->text
Pricing:
Introducing Mistral Large 2 2411, the latest iteration of the Mistral Large 2 model, launched alongside Pixtral Large 2411. This update represents a substantial enhancement over its predecessor, Mistral Large 24.07, delivering significant advancements in long-context comprehension, a refined system prompt, and improved accuracy in function calling. Designed to elevate performance and precision, Mistral Large 2 2411 sets a new standard for intelligent language models.
Mistral: Pixtral Large 2411
- Context Length:
- 128,000 tokens
- Architecture:
- text+image->text
Pricing:
Pixtral Large is a cutting-edge, open-weight multimodal model with 124 billion parameters, built upon the robust foundation of Mistral Large 2. This advanced model excels at interpreting a wide range of data types, including documents, charts, and natural images, making it a versatile tool for complex tasks.
Pixtral Large is available under two distinct licenses: the Mistral Research License (MRL) for academic and research purposes, and the Mistral Commercial License, which allows for experimentation, testing, and deployment in commercial environments. This dual licensing structure ensures flexibility for both educational and business applications.
Mistral Medium
- Context Length:
- 32,000 tokens
- Architecture:
- text->text
Pricing:
Introducing Mistral AI's cutting-edge, medium-sized model—a proprietary solution driven by a closed-source prototype. This advanced model delivers exceptional performance in reasoning, coding, JSON handling, and conversational capabilities. Rigorously tested, it stands toe-to-toe with leading flagship models from industry competitors, showcasing its versatility and robust capabilities across a wide range of tasks.
Ready to build with Mistral?
Start using these powerful models in your applications with our flexible pricing plans.