bert-base-nli-stsb-mean-tokens

Maintained By
sentence-transformers

bert-base-nli-stsb-mean-tokens

PropertyValue
Authorsentence-transformers
Vector Dimension768
PaperSentence-BERT: Sentence Embeddings using Siamese BERT-Networks
StatusDeprecated

What is bert-base-nli-stsb-mean-tokens?

This is a deprecated sentence transformer model that maps sentences and paragraphs to 768-dimensional dense vector spaces. While historically used for tasks like clustering and semantic search, it is no longer recommended due to its relatively low-quality embeddings.

Implementation Details

The model architecture consists of a BERT-based transformer followed by a pooling layer. It uses mean pooling over token embeddings and can process sequences up to 128 tokens. The model can be implemented using either the sentence-transformers library or HuggingFace Transformers.

  • Built on BERT-base architecture
  • Uses mean pooling strategy
  • Maximum sequence length of 128 tokens
  • Implements Siamese network architecture

Core Capabilities

  • Sentence and paragraph embedding generation
  • Semantic similarity comparison
  • Text clustering
  • Semantic search operations

Frequently Asked Questions

Q: What makes this model unique?

This model was one of the early implementations of Sentence-BERT architecture, using mean token pooling strategy. However, it has been superseded by more modern alternatives available on SBERT.net.

Q: What are the recommended use cases?

Due to its deprecated status, it is NOT recommended for any production use cases. Users should instead refer to the newer models listed on SBERT.net for better performance in sentence embedding tasks.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.