Mistral Large
Mistral Large overview
#TL;DR(中文)
- 是 Mistral AI 的旗舰code
Mistral Large之一,定位是更强的 multilingual/reasoning/coding 能力,并支持较长 context window。codeLLM - 适合用作 “capability model”:复杂分析、长文档理解、以及需要更强 /JSON 的场景。code
Tool Calling - 工程选型建议:对你的任务做 (准确性、格式稳定性、长上下文 recall),并关注 cost/latency 的实际表现。code
evaluation
#中文导读(术语保留英文)
看这页时建议重点关注:
- context window(32K)与长上下文任务的可靠性
- function calling / JSON 支持(是否适合 agentic workflow)
- multilingual 能力与具体语言覆盖
#Original (English)
Mistral AI releases Mistral, their most advanced large language model (LLM) with strong multilingual, reasoning, maths, and code generation capabilities. Mistral Large is made available through Mistral platform called la Plataforme and Microsoft Azure. It's also available to test in their new chat app, le Chat.
Below is a chart showing how Mistral Large compares with other powerful LLMs like GPT-4 and Gemini Pro. It ranks second next to GPT-4 on the MMLU benchmark with a score of 81.2%.

#Mistral Large Capabilities
Mistral Large's capabilities and strengths include:
- 32K tokens context window
- has native multilingual capacities (fluent in English, French, Spanish, German, and Italian)
- strong capabilities in reasoning, knowledge, maths, and coding benchmarks
- function calling and JSON format natively supported
- a low-latency model called Mistral Small was also released
- allows developers to design moderation policies with its precise instruction-following
#Reasoning and Knowledge
The table below shows how Mistral Large performs on common reasoning and knowledge benchmarks. It largely falls behind GPT-4 but it's the superior model compared to other LLMs like Claude 2 and Gemini Pro 1.0.

#Maths & Code Generation
The table below shows how Mistral Large performs on common maths and coding benchmarks. Mistral Large demonstrates strong performance on the Math and GSM8K benchmarks but it is significantly outperformed on coding benchmarks by models like Gemini Pro and GPT-4.

#Multilinguality
The table below demonstrates Mistral Large performance on multilingual reasoning benchmarks. Mistral Large outperforms Mixtral 8x7B and Llama 2 70B in all languages, including French, German, Spanish, and Italian.

#Mistral Small
In addition to the release of Mistral Large, a smaller model and optimized model called Mistral Small is also announced. Mistral Small is optimized for low-latency workloads and outperforms Mixtral 8x7B. Mistral AI reports that this model has strong capacities around RAG-enablement, function calling, and JSON format.
#Mistral Endpoints and Model Selection
Here is a list of all the endpoints provided by Mistral AI.
Mistral AI has also published a comprehensive guide on better model selection when considering performance and cost trade-offs.
Figures source: https://mistral.ai/news/mistral-large/