bleurt-base-512

bleurt-base-512

Elron

PyTorch implementation of BLEURT for text evaluation metrics, based on BERT architecture. Supports reference-candidate comparison with 512 token limit.

PropertyValue
AuthorElron
FrameworkPyTorch
Base ArchitectureBERT
Downloads67,737
PaperBLEURT: Learning Robust Metrics for Text Generation

What is bleurt-base-512?

BLEURT-base-512 is a PyTorch implementation of the original BLEURT model, designed for evaluating text generation quality. This model serves as a learned metric for comparing reference texts with candidate generations, supporting sequences up to 512 tokens in length.

Implementation Details

The model is implemented as a sequence classification task using the Transformers library. It takes pairs of reference and candidate texts as input and outputs similarity scores. The implementation is based on the original Google Research work, converted from TensorFlow to PyTorch for broader accessibility.

  • Built on BERT architecture
  • Supports 512 token sequence length
  • Provides numerical similarity scores
  • Implemented using HuggingFace Transformers

Core Capabilities

  • Text similarity scoring
  • Generation quality assessment
  • Reference-based evaluation
  • Batch processing support

Frequently Asked Questions

Q: What makes this model unique?

BLEURT-base-512 stands out for its learned approach to text evaluation, moving beyond traditional metrics like BLEU or ROUGE. It provides more nuanced similarity assessments by leveraging BERT's contextual understanding.

Q: What are the recommended use cases?

The model is ideal for evaluating machine translation outputs, text summarization results, and any scenario requiring quality assessment of generated text against reference texts. It's particularly useful in research and development of text generation systems.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026