snowflake-arctic-embed-m-v2.0

snowflake-arctic-embed-m-v2.0

Snowflake

Multilingual embedding model with 305M parameters optimized for retrieval, supporting 128-byte compression and 8192 context window, ideal for enterprise search.

PropertyValue
Total Parameters305M
Non-embedding Parameters113M
Embedding Dimensions768
Context Window8192 tokens
LicenseApache 2.0
Model URLhttps://huggingface.co/Snowflake/snowflake-arctic-embed-m-v2.0

What is snowflake-arctic-embed-m-v2.0?

Snowflake's Arctic-embed-m-v2.0 is a state-of-the-art multilingual embedding model designed for enterprise-grade text retrieval. It represents a significant advancement in multilingual AI, offering superior performance across both English and non-English content without compromising quality in either domain.

Implementation Details

The model utilizes Matryoshka Representation Learning (MRL) and quantization-aware embedding training to achieve highly efficient compression capabilities. It can maintain high-quality retrieval even with embeddings as small as 128 bytes per vector, making it highly efficient for large-scale deployments. The architecture incorporates RoPE (Rotary Position Embedding) to support an extended context window of 8192 tokens.

  • Benchmark Performance: Achieves 55.4 on BEIR(15), 55.2 on MIRACL(4), and 53.9 on CLEF(Full)
  • Vector Compression: Only 3% quality degradation with 3x size reduction
  • Efficient Architecture: 113M non-embedding parameters for fast inference

Core Capabilities

  • High-quality multilingual text retrieval
  • Efficient compression without significant performance loss
  • Extended context window support
  • Enterprise-grade performance at scale
  • Seamless integration with popular frameworks like Sentence Transformers and Hugging Face

Frequently Asked Questions

Q: What makes this model unique?

The model's ability to excel in both English and non-English retrieval while maintaining competitive performance across multiple benchmarks sets it apart. Its compression capabilities and extended context window make it particularly suitable for enterprise applications.

Q: What are the recommended use cases?

The model is ideal for multilingual search systems, document retrieval applications, and large-scale enterprise search implementations where efficiency and accuracy across multiple languages are crucial.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026