ms-marco-MiniLM-L6-v2

Maintained By
cross-encoder

ms-marco-MiniLM-L6-v2

PropertyValue
Authorcross-encoder
Model TypeCross-Encoder
Performance (NDCG@10)74.30
Speed1800 docs/sec on V100 GPU
Model HubHugging Face

What is ms-marco-MiniLM-L6-v2?

ms-marco-MiniLM-L6-v2 is a specialized cross-encoder model designed for passage ranking tasks, particularly optimized for the MS Marco dataset. It represents version 2 of the architecture, offering significant improvements over its predecessors in both performance and efficiency.

Implementation Details

The model implements a cross-encoder architecture with 6 layers (L6) of MiniLM, striking an optimal balance between computational efficiency and performance. It can be easily integrated using either SentenceTransformers or the Transformers library, supporting both high-level and low-level implementations.

  • Achieves 74.30 NDCG@10 on TREC DL 19
  • Processes 1800 documents per second on V100 GPU
  • Delivers 39.01 MRR@10 on MS Marco Dev set

Core Capabilities

  • Passage ranking and reranking
  • Query-passage relevance scoring
  • Information retrieval enhancement
  • Compatible with major deep learning frameworks

Frequently Asked Questions

Q: What makes this model unique?

This model offers an excellent trade-off between performance and speed, outperforming many larger models while maintaining reasonable computational requirements. Its 6-layer architecture proves optimal for most passage ranking tasks.

Q: What are the recommended use cases?

The model excels in information retrieval scenarios where ranking or reranking of passages is required. It's particularly effective when used in conjunction with initial retrieval systems like ElasticSearch for re-ranking top results.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.