sbert-all-MiniLM-L6-with-pooler

Maintained By
optimum

SBERT-all-MiniLM-L6-with-pooler

PropertyValue
Authoroptimum
Vector Dimension384
Base ModelMiniLM-L6-H384-uncased
Training Data1B+ sentence pairs
Max Sequence Length256 tokens

What is sbert-all-MiniLM-L6-with-pooler?

SBERT-all-MiniLM-L6-with-pooler is an ONNX-optimized sentence embedding model that maps text to dense 384-dimensional vectors. Unlike standard SBERT models, this version includes both last_hidden_state and pooler_output, making it more versatile for downstream tasks. The model was trained on over 1 billion sentence pairs using contrastive learning, making it particularly effective for semantic search, clustering, and similarity tasks.

Implementation Details

The model is built on MiniLM-L6-H384-uncased and fine-tuned using a contrastive learning objective. Training utilized TPU v3-8 hardware with a batch size of 1024 over 100k steps. The implementation uses the AdamW optimizer with a 2e-5 learning rate and includes a 500-step warmup period.

  • ONNX-optimized for efficient inference
  • Dual output architecture (hidden states and pooler)
  • Trained on diverse datasets including Reddit, WikiAnswers, and academic citations
  • Optimized for sentences and short paragraphs up to 256 tokens

Core Capabilities

  • Semantic text embedding generation
  • Information retrieval and clustering
  • Sentence similarity computation
  • Cross-lingual text comparison
  • Document classification support

Frequently Asked Questions

Q: What makes this model unique?

This model stands out due to its ONNX optimization and dual output architecture, providing both hidden states and pooler outputs. It's trained on an exceptionally large and diverse dataset of over 1 billion sentence pairs, making it robust for various text similarity tasks.

Q: What are the recommended use cases?

The model excels in semantic search applications, document clustering, similarity analysis, and information retrieval tasks. It's particularly effective for short to medium-length text processing, with optimal performance on content under 256 tokens.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.