col-minilm

Maintained By
vespa-engine

col-minilm

PropertyValue
Parameters22.3M
LicenseMIT
Base Modelcross-encoder/ms-marco-MiniLM-L-6-v2
PaperColBERT Paper

What is col-minilm?

col-minilm is an efficient passage ranking model based on the ColBERT architecture, specifically optimized for use with Vespa.ai. It's built upon the MS-Marco-MiniLM-L-6-v2 base model and trained using the ColBERT training routine, offering a perfect balance between performance and efficiency.

Implementation Details

The model employs a contextualized late interaction approach over BERT, making it approximately 2x faster than its medium counterpart while maintaining competitive performance. It achieves an impressive MRR@10 of 0.364 on the MS Marco dev set, with strong recall metrics across different k values.

  • Recall@50: 0.816
  • Recall@200: 0.905
  • Recall@1000: 0.939
  • ONNX export support for Vespa.ai integration

Core Capabilities

  • Efficient passage search and ranking
  • Optimized for production deployment with Vespa.ai
  • Contextual embedding generation
  • Support for both query and document encoding

Frequently Asked Questions

Q: What makes this model unique?

The model's distinctive feature is its efficient architecture that maintains high-quality results while being significantly faster than larger alternatives. It's specifically optimized for production deployment with Vespa.ai and includes ONNX export capabilities.

Q: What are the recommended use cases?

This model is ideal for production-scale passage ranking tasks, particularly when integrated with Vespa.ai. It's especially suitable for applications requiring efficient search and ranking over large document collections while maintaining high accuracy.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.