gbert-large-zeroshot-nli

Maintained By
svalabs

GBERT-Large-ZeroShot-NLI

PropertyValue
Base ModelGBERT-large
TaskZero-shot Classification / NLI
Training Data847,862 translated pairs (MNLI, ANLI, SNLI)
XNLI Accuracy85.6%
Model HubHugging Face

What is gbert-large-zeroshot-nli?

GBERT-Large-ZeroShot-NLI is a specialized German language model developed by SVALabs, built on deepset.ai's German BERT large architecture. The model is specifically fine-tuned for zero-shot classification tasks using Natural Language Inference (NLI) training. It leverages a massive dataset of 847,862 machine-translated sentence pairs from MNLI, ANLI, and SNLI datasets.

Implementation Details

The model implements a zero-shot classification approach through NLI training, allowing it to categorize text without specific training examples for each category. It excels particularly in German text classification tasks, demonstrating superior performance compared to other German and multilingual models.

  • Built on GBERT-large architecture
  • Fine-tuned on translated NLI datasets
  • Achieves 81% accuracy on 10kGNAD dataset classification
  • Supports flexible hypothesis templates for different classification scenarios

Core Capabilities

  • Zero-shot text classification in German
  • Natural Language Inference tasks
  • Flexible hypothesis template support
  • High performance on general classification tasks

Frequently Asked Questions

Q: What makes this model unique?

The model stands out for its exceptional performance in German language zero-shot classification, achieving significantly better results (81% accuracy) compared to other German language models in benchmark tests. It's specifically optimized for German text analysis with carefully translated training data.

Q: What are the recommended use cases?

The model is ideal for German text classification tasks where predefined training data isn't available. It performs particularly well with single-word labels using the template "In diesem Satz geht es um das Thema {}" and can handle more complex classifications using templates like "Weil {}" or "Daher {}".

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.