gbert-large-zeroshot-nli

gbert-large-zeroshot-nli

svalabs

German BERT-large model fine-tuned for zero-shot classification using NLI, achieving 85.6% accuracy on XNLI test set. Trained on 847K translated pairs.

PropertyValue
Base ModelGBERT-large
TaskZero-shot Classification / NLI
Training Data847,862 translated pairs (MNLI, ANLI, SNLI)
XNLI Accuracy85.6%
Model HubHugging Face

What is gbert-large-zeroshot-nli?

GBERT-Large-ZeroShot-NLI is a specialized German language model developed by SVALabs, built on deepset.ai's German BERT large architecture. The model is specifically fine-tuned for zero-shot classification tasks using Natural Language Inference (NLI) training. It leverages a massive dataset of 847,862 machine-translated sentence pairs from MNLI, ANLI, and SNLI datasets.

Implementation Details

The model implements a zero-shot classification approach through NLI training, allowing it to categorize text without specific training examples for each category. It excels particularly in German text classification tasks, demonstrating superior performance compared to other German and multilingual models.

  • Built on GBERT-large architecture
  • Fine-tuned on translated NLI datasets
  • Achieves 81% accuracy on 10kGNAD dataset classification
  • Supports flexible hypothesis templates for different classification scenarios

Core Capabilities

  • Zero-shot text classification in German
  • Natural Language Inference tasks
  • Flexible hypothesis template support
  • High performance on general classification tasks

Frequently Asked Questions

Q: What makes this model unique?

The model stands out for its exceptional performance in German language zero-shot classification, achieving significantly better results (81% accuracy) compared to other German language models in benchmark tests. It's specifically optimized for German text analysis with carefully translated training data.

Q: What are the recommended use cases?

The model is ideal for German text classification tasks where predefined training data isn't available. It performs particularly well with single-word labels using the template "In diesem Satz geht es um das Thema {}" and can handle more complex classifications using templates like "Weil {}" or "Daher {}".

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026