Get started with Mistral models in a few clicks via our developer platform hosted on Mistral’s infrastructure and build your own applications and services. Our servers are hosted in EU.
We release the world’s most capable open models, enabling frontier AI innovation.
Our portable developer platform serves our open and optimized models for building fast and intelligent applications. We offer flexible access options!
We’re committed to empower the AI community with open technology. Our open models sets the bar for efficiency, and are available for free under Apache 2.0, a fully permissive license, that allows to use the models anywhere without any restriction.
Our very first. A 7B transformer model, fast-deployed and easily customisable. Small, yet very powerful for a variety of use cases.
A 7B sparse Mixture-of-Experts (SMoE). Uses 12.9B active parameters out of 45B total.
Mixtral 8x22B is currently the most performant open model. A 22B sparse Mixture-of-Experts (SMoE). Uses only 39B active parameters out of 141B.
Our optimized commercial models are designed for performance and are available via our flexible deployment options.
Cost-efficient reasoning for low-latency workloads.
Top-tier reasoning for high-complexity tasks. The most powerful model of the Mistral AI family.
State-of-the-art semantic for extracting representation of text extracts.
We’re constantly innovating to provide the most capable and efficient models.
We build models that offer unparalleled cost efficiency for their respective sizes, delivering the best performance-to-cost ratio available on the market. Mixtral 8x22B is the most powerful open source model with significantly fewer parameters than its competition:
Our open models are truly open source, licensed under Apache 2.0, a fully permissive license that allows for unrestricted use in any context.
Access our latest products via our developer platform, hosted in Europe
from mistralai.client import MistralClient
from mistralai.models.chat_completion import ChatMessage
api_key = os.environ["MISTRAL_API_KEY"]
model = "mistral-tiny"
client = MistralClient(api_key=api_key)
messages = [
ChatMessage(role="user",
content="Who is the most renowned French painter?")
]
La Plateforme is developers’ preferred way to access all Mistral Al’s models. Hosted and served on Mistral Al infrastructure, in Europe.
from mistralai.client import MistralClient
from mistralai.models.chat_completion import ChatMessage
api_key = os.environ["MISTRAL_API_KEY"]
model = "mistral-tiny"
client = MistralClient(api_key=api_key)
messages = [
ChatMessage(role="user",
content="Who is the most renowned French painter?")
]
Input | Output | ||
---|---|---|---|
open-mistral-7b | A 7B transformer model, fast-deployed and easily customisable. | $0.25 /1M tokens | $0.25 /1M tokens |
open-mixtral-8x7b | A 7B sparse Mixture-of-Experts (SMoE). Uses 12.9B active parameters out of 45B total. | $0.7 /1M tokens | $0.7 /1M tokens |
open-mixtral-8x22b | Mixtral 8x22B is currently the most performant open model. A 22B sparse Mixture-of-Experts (SMoE). Uses only 39B active parameters out of 141B. | $2 /1M tokens | $6 /1M tokens |
Input | Output | ||
---|---|---|---|
open-mistral-7b | A 7B transformer model, fast-deployed and easily customisable. | 0.2€ /1M tokens | 0.2€ /1M tokens |
open-mixtral-8x7b | A 7B sparse Mixture-of-Experts (SMoE). Uses 12.9B active parameters out of 45B total. | 0.65€ /1M tokens | 0.65€ /1M tokens |
open-mixtral-8x22b | Mixtral 8x22B is currently the most performant open model. A 22B sparse Mixture-of-Experts (SMoE). Uses only 39B active parameters out of 141B. | 1.9€ /1M tokens | 5.6€ /1M tokens |
Input | Output | ||
---|---|---|---|
mistral-small | Cost-efficient reasoning for low-latency workloads. | $1 /1M tokens | $3 /1M tokens |
mistral-medium | Will soon be deprecated | $2.7 /1M tokens | $8.1 /1M tokens |
mistral-large | Top-tier reasoning for high-complexity tasks. The most powerful model of the Mistral AI family. | $4 /1M tokens | $12 /1M tokens |
Input | Output | ||
---|---|---|---|
mistral-small | Cost-efficient reasoning for low-latency workloads. | 0.9€ /1M tokens | 2.8€ /1M tokens |
mistral-medium | Will soon be deprecated | 2.5€ /1M tokens | 7.5€ /1M tokens |
mistral-large | Top-tier reasoning for high-complexity tasks. The most powerful model of the Mistral AI family. | 3.8€ /1M tokens | 11.3€ /1M tokens |
Input | Output | ||
---|---|---|---|
mistral-embed | State-of-the-art semantic for extracting representation of text extracts. | $0.1 /1M tokens | $0.1 /1M tokens |
Input | Output | ||
---|---|---|---|
mistral-embed | State-of-the-art semantic for extracting representation of text extracts. | 0.1€ /1M tokens | 0.1€ /1M tokens |