sbert-all-MiniLM-L6-with-pooler

sbert-all-MiniLM-L6-with-pooler

optimum

Sentence embedding model that converts text to 384-dimensional vectors, optimized for ONNX runtime with both hidden state and pooler outputs

PropertyValue
Authoroptimum
Vector Dimension384
Base ModelMiniLM-L6-H384-uncased
Training Data1B+ sentence pairs
Max Sequence Length256 tokens

What is sbert-all-MiniLM-L6-with-pooler?

SBERT-all-MiniLM-L6-with-pooler is an ONNX-optimized sentence embedding model that maps text to dense 384-dimensional vectors. Unlike standard SBERT models, this version includes both last_hidden_state and pooler_output, making it more versatile for downstream tasks. The model was trained on over 1 billion sentence pairs using contrastive learning, making it particularly effective for semantic search, clustering, and similarity tasks.

Implementation Details

The model is built on MiniLM-L6-H384-uncased and fine-tuned using a contrastive learning objective. Training utilized TPU v3-8 hardware with a batch size of 1024 over 100k steps. The implementation uses the AdamW optimizer with a 2e-5 learning rate and includes a 500-step warmup period.

  • ONNX-optimized for efficient inference
  • Dual output architecture (hidden states and pooler)
  • Trained on diverse datasets including Reddit, WikiAnswers, and academic citations
  • Optimized for sentences and short paragraphs up to 256 tokens

Core Capabilities

  • Semantic text embedding generation
  • Information retrieval and clustering
  • Sentence similarity computation
  • Cross-lingual text comparison
  • Document classification support

Frequently Asked Questions

Q: What makes this model unique?

This model stands out due to its ONNX optimization and dual output architecture, providing both hidden states and pooler outputs. It's trained on an exceptionally large and diverse dataset of over 1 billion sentence pairs, making it robust for various text similarity tasks.

Q: What are the recommended use cases?

The model excels in semantic search applications, document clustering, similarity analysis, and information retrieval tasks. It's particularly effective for short to medium-length text processing, with optimal performance on content under 256 tokens.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026