Mistral-Large-Instruct-2411-exl2

Maintained By
bartowski

Mistral-Large-Instruct-2411-exl2

PropertyValue
Base Modelmistralai/Mistral-Large-Instruct-2411
LicenseMistral Research License (MRL)
Supported Languages10 (English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Russian, Korean)
Quantization FrameworkExLlama v2

What is Mistral-Large-Instruct-2411-exl2?

This is a quantized version of the Mistral-Large-Instruct model, optimized using ExLlama v2 framework. The model offers multiple quantization options ranging from 2.2 to 6.5 bits per weight, allowing users to balance between model size and performance based on their specific needs. The quantization was performed using turboderp's ExLlamaV2 v0.2.4 with default calibration datasets.

Implementation Details

The model implements different quantization levels through separate branches, each offering a different compression ratio:

  • 6.5 bits per weight (with 8-bit lm_head layer)
  • 5.0 bits per weight
  • 4.25 bits per weight
  • 3.75 bits per weight
  • 3.5 bits per weight
  • 3.0 bits per weight
  • 2.2 bits per weight

Core Capabilities

  • Multi-language support across 10 major languages
  • Flexible quantization options for different deployment scenarios
  • Maintained text generation quality with reduced model size
  • Optimized for research purposes

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its variety of quantization options while maintaining the core capabilities of the original Mistral-Large-Instruct model. It's specifically optimized for research purposes and offers a range of compression levels to suit different hardware constraints.

Q: What are the recommended use cases?

The model is strictly limited to non-commercial research purposes as per the Mistral Research License. It's ideal for academic research, personal scientific exploration, and non-profit applications. Commercial use requires explicit permission from Mistral AI.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.