bart-large

bart-large

facebook

BART-large: A powerful transformer-based seq2seq model by Facebook, pre-trained for text generation and comprehension tasks with 149K+ downloads

PropertyValue
AuthorFacebook
LicenseApache-2.0
PaperView Paper
Downloads149,763
LanguageEnglish

What is bart-large?

BART-large is a powerful transformer-based sequence-to-sequence model developed by Facebook. It combines a bidirectional BERT-like encoder with an autoregressive GPT-like decoder, making it particularly effective for both text generation and comprehension tasks. The model is pre-trained using a unique denoising approach where text is first corrupted and then reconstructed.

Implementation Details

The model implements a sophisticated architecture that leverages both bidirectional encoding and autoregressive decoding. It can be easily integrated using the Hugging Face transformers library, supporting PyTorch implementations.

  • Transformer-based encoder-decoder architecture
  • Bidirectional encoding similar to BERT
  • Autoregressive decoding like GPT
  • Pre-trained on English language corpus

Core Capabilities

  • Text generation and summarization
  • Machine translation
  • Text comprehension and classification
  • Question answering
  • Text infilling capabilities

Frequently Asked Questions

Q: What makes this model unique?

BART's uniqueness lies in its denoising pre-training approach and its hybrid architecture that combines the best aspects of BERT and GPT. This makes it particularly versatile for both generation and comprehension tasks.

Q: What are the recommended use cases?

While the base model can be used for text infilling, it's primarily designed to be fine-tuned for specific tasks. It excels in text generation tasks like summarization and translation, as well as comprehension tasks like classification and question answering.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026