Mistral-7B-Instruct-v0.1 is an instruction-tuned model based on Mistral-7B-v0.1, designed for improved performance in conversational AI and assistant tasks. It provides robust handling of instruction-following prompts in dialogue. This model is accessible through both the Mistral and Hugging Face Transformers libraries.
Developed by Mistral AI, with contributions from a diverse team including Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, and others.
Mistral-7B-Instruct-v0.1 is suitable for:
This model lacks built-in moderation mechanisms and should not be deployed in scenarios that require strict content filtering.
Mistral-7B-Instruct-v0.1 is based on the Mistral-7B architecture, with the following design features:
Mistral-7B-Instruct-v0.1 is fine-tuned using various publicly available datasets, curated to enhance the model’s instruction-following and dialogue generation capabilities. For comprehensive details, refer to the release paper and blog post.
Inference can be performed via both Mistral’s proprietary library and Hugging Face Transformers. Mistral’s framework supports low-latency deployments.
The model currently does not include moderation mechanisms. Users should exercise caution when deploying Mistral-7B-Instruct-v0.1 in applications that require strict content safety and moderation.
Mistral AI invites community contributions, particularly for enhancing the alignment between Mistral’s tokenizer and Transformers. Contributions, including pull requests to refine the model, are encouraged.
KeyError
transformers-v4.33.4
If you use Mistral-7B-Instruct-v0.1 in your research, please cite:
@misc{mistralai2024mistral7b, author = {Mistral AI}, title = {Mistral-7B-Instruct-v0.1: Fine-tuned Large Language Model for Instruction Following}, year = {2024}, url = {https://github.com/mistralai/mistral-models}, publisher = {Mistral AI} }