German_Semantic_STS_V2

German_Semantic_STS_V2

aari1995

A German language semantic similarity model with 336M parameters, achieving SOTA performance on STS benchmarks with 86.26% Spearman correlation.

PropertyValue
Parameter Count336M
Model TypeSentence Transformer
ArchitecturegBERT-large based with mean pooling
Dimension1024

What is German_Semantic_STS_V2?

German_Semantic_STS_V2 is a state-of-the-art German language model specifically designed for semantic similarity tasks. Built on gBERT-large architecture, it maps sentences and paragraphs to 1024-dimensional dense vector space, achieving an impressive 86.26% Spearman correlation on German STS benchmarks.

Implementation Details

The model uses a sentence-transformers framework with a two-component architecture: a transformer module with 512 max sequence length and a pooling layer implementing mean token pooling. It was trained using ContrastiveLoss with cosine distance metric and a margin of 0.5.

  • Trained with AdamW optimizer (lr=5e-06)
  • 4 epochs with 576 warmup steps
  • Batch size of 4 with gradient clipping at 1.0

Core Capabilities

  • Semantic text similarity scoring
  • Clustering of German text
  • Semantic search applications
  • Cross-lingual tasks (particularly German-English)
  • Domain and intent classification

Frequently Asked Questions

Q: What makes this model unique?

This model outperforms previous German language models on semantic similarity tasks, achieving higher Spearman correlation (86.26%) than established multilingual models like XLM-R and other German-specific models.

Q: What are the recommended use cases?

The model excels in semantic search, document clustering, and similarity analysis for German text. It's particularly effective for tasks requiring nuanced understanding of semantic relationships between texts.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026