This work introduces Mistral 7B v0.1, a 7-billion-parameter language model engineered for superior performance and efficiency, which leverages grouped-query attention (GQA) for faster inference, coupled with sliding window attention (SWA) to effectively handle sequences of arbitrary length with a reduced inference cost.
We introduce Mistral 7B v0.1, a 7-billion-parameter language model engineered for superior performance and efficiency. Mistral 7B outperforms Llama 2 13B across all evaluated benchmarks, and Llama 1 34B in reasoning, mathematics, and code generation. Our model leverages grouped-query attention (GQA) for faster inference, coupled with sliding window attention (SWA) to effectively handle sequences of arbitrary length with a reduced inference cost. We also provide a model fine-tuned to follow instructions, Mistral 7B -- Instruct, that surpasses the Llama 2 13B -- Chat model both on human and automated benchmarks. Our models are released under the Apache 2.0 license.
Diego de Las Casas
5 papers
M. Lachaux
8 papers
Thomas Wang
4 papers
Teven Le Scao
6 papers
Timothée Lacroix
6 papers
Albert Qiaochu Jiang
2 papers
Alexandre Sablayrolles
2 papers
Chris Bamford
2 papers
Devendra Singh Chaplot
6 papers
Florian Bressand
2 papers
Gianna Lengyel
2 papers
Lucile Saulnier
3 papers
Lélio Renard Lavaud
2 papers
Pierre Stock
2 papers
William El Sayed
2 papers