Skip to main content

Title

Add Mistral-7B-Instruct-v0.2 to KAITO supported model list.

Glossary

N/A

Summary

  • Model description: Launched last September, Mistral-7B has quickly become a standout among large language models (LLMs) with its 7.3B parameters. It surpasses models nearly twice its size in benchmarks, showcasing its efficiency and cost-effectiveness. Mistral has also forged strategic alliances with major cloud platforms including Azure, AWS, and GCP. For more information, refer to the Mistral-7B Documentation and access the model on Hugging Face. The newly introduced Mistral-7B-Instruct-v0.1 is an instruct fine-tuned iteration of the original, optimized for understanding and executing specific instructions. This enhances mistral's utility in conversational applications.
  • Model usage statistics: In the past month, Mistral-7B-Instruct-v0.1 has garnered 424,580 downloads on Hugging Face, reflecting its widespread popularity. Google Trends data shows a high level of search interest in "mistral ai 7b", indicating strong market curiosity.
  • Model license: Mistral-7B-Instruct-v0.1 is distributed under the Apache 2.0 license, ensuring broad usability and modification rights.

Requirements

The following table describes the basic model characteristics and the resource requirements of running it.

FieldNotes
Family nameMistral
Typetext generation
Download sitehttps://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2
Versionb70aa86578567ba3301b21c8a27bea4e8f6d6d61
Storage size50GB
GPU count1
Total GPU memory16GB
Per GPU memoryN/A

Runtimes

This section describes how to configure the runtime framework to support the inference calls.

OptionsNotes
RuntimeHuggingface Transformer
Distributed InferenceFalse
Custom configurationsPrecision: BF16. Can run on one machine with total of 16GB of GPU Memory.

History

  • 02/05/2024: Open proposal PR.
  • 02/06/2024: Update proposal PR - Mistral-7B-Instruct-v0.2