gbert-base

gbert-base

deepset

German BERT language model with 111M parameters, trained on Wikipedia, OPUS & OpenLegalData. Optimized for German NLP tasks with MIT license.

PropertyValue
Parameter Count111M parameters
LicenseMIT
Research PaperView Paper
Training DataWikipedia, OPUS, OpenLegalData

What is gbert-base?

gbert-base is a German BERT language model released in October 2020, representing a collaborative effort between the creators of the original German BERT and dbmdz BERT. This model features 111M parameters and has been specifically optimized for German language processing tasks.

Implementation Details

The model implements the BERT base architecture and has been trained on a diverse dataset including German Wikipedia, OPUS, and OpenLegalData. It demonstrates impressive performance metrics, achieving 78.17% on GermEval18 Coarse, 50.90% on GermEval18 Fine, and 87.98% on GermEval14 tasks.

  • Built on BERT base architecture
  • Optimized for German language understanding
  • Trained on comprehensive German text corpora
  • Supports PyTorch and TensorFlow frameworks

Core Capabilities

  • Fill-Mask task operations
  • German text understanding and processing
  • Compatible with Transformer-based applications
  • Supports inference endpoints for production deployment

Frequently Asked Questions

Q: What makes this model unique?

This model represents a collaborative improvement over previous German BERT models, showing superior performance on various benchmarks while maintaining a practical base model size.

Q: What are the recommended use cases?

The model is ideal for German language processing tasks, including text classification, named entity recognition, and masked language modeling. It's particularly well-suited for applications requiring deep understanding of German text.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026