distilbart-cnn-12-6

distilbart-cnn-12-6

sshleifer

A distilled BART model trained on CNN/DailyMail dataset for text summarization, offering 1.24x speedup with comparable ROUGE scores to BART-large-cnn

PropertyValue
Parameters306M
LicenseApache 2.0
ArchitectureBART Distilled
Training DataCNN/DailyMail Dataset

What is distilbart-cnn-12-6?

DistilBART CNN 12-6 is a compressed version of the BART-large-cnn model, designed for efficient text summarization. It maintains nearly identical performance to its larger counterpart while offering improved inference speed. The model name indicates its architecture: 12 encoder layers and 6 decoder layers.

Implementation Details

This model is implemented using the BartForConditionalGeneration architecture and achieves a 1.24x speedup compared to the baseline BART-large-cnn model. With 306M parameters, it achieves impressive ROUGE scores (ROUGE-2: 21.26, ROUGE-L: 30.59) that are comparable to the full model.

  • Optimized inference time of 307ms
  • Maintains 99% of the original model's performance
  • Designed for production deployment with reduced computational requirements

Core Capabilities

  • Text summarization optimized for news articles
  • Efficient processing of long documents
  • Compatible with both PyTorch and JAX frameworks
  • Suitable for deployment on inference endpoints

Frequently Asked Questions

Q: What makes this model unique?

This model achieves an optimal balance between performance and efficiency, using knowledge distillation to compress BART while maintaining high-quality summarization capabilities.

Q: What are the recommended use cases?

The model is particularly well-suited for news article summarization, content condensation, and production environments where efficiency is crucial without compromising quality.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026