ms-marco-MiniLM-L4-v2

Maintained By
cross-encoder

ms-marco-MiniLM-L4-v2

PropertyValue
Model TypeCross-Encoder
Authorcross-encoder
Performance (NDCG@10)73.04
Processing Speed2500 docs/sec
Model URLHuggingFace

What is ms-marco-MiniLM-L4-v2?

ms-marco-MiniLM-L4-v2 is a specialized cross-encoder model designed for passage ranking tasks, particularly optimized for the MS Marco dataset. It represents the second version of the 4-layer MiniLM architecture, striking an excellent balance between performance and efficiency.

Implementation Details

The model implements a cross-encoder architecture specifically trained on the MS Marco Passage Ranking task. It can be easily integrated using either SentenceTransformers or the Transformers library, making it versatile for different implementation needs.

  • Built on MiniLM architecture with 4 layers
  • Achieves 37.70 MRR@10 on MS Marco Dev set
  • Processes 2500 documents per second on V100 GPU
  • Supports both query-passage pair scoring and ranking

Core Capabilities

  • Efficient passage ranking and retrieval
  • Query-passage relevance scoring
  • Information retrieval enhancement
  • Compatible with major deep learning frameworks

Frequently Asked Questions

Q: What makes this model unique?

This model offers an optimal balance between performance and speed, achieving a strong NDCG@10 score of 73.04 while maintaining relatively high processing speed. It's particularly effective for production environments where both accuracy and efficiency are crucial.

Q: What are the recommended use cases?

The model is ideal for information retrieval systems, particularly when implementing a retrieve & re-rank architecture. It's well-suited for applications requiring passage ranking, document search, and query-passage relevance scoring.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.