led-base-book-summary

led-base-book-summary

pszemraj

LED-based model (162M params) for long-form text summarization. Handles 16K tokens, optimized for books/technical content, ROUGE-1: 33.45

PropertyValue
Parameter Count162M
Model TypeLongformer Encoder-Decoder
LicenseBSD-3-Clause
Max Input Length16,384 tokens
ROUGE-1 Score33.45

What is led-base-book-summary?

LED-base-book-summary is a specialized text summarization model built on the Longformer Encoder-Decoder architecture. Fine-tuned on the BookSum dataset, this model excels at condensing long-form content while maintaining coherent narrative structure. With 162M parameters, it offers an efficient balance between computational requirements and summary quality.

Implementation Details

The model utilizes a sophisticated architecture that enables processing of up to 16,384 tokens per batch, making it ideal for lengthy documents. It was trained for 16 epochs with carefully optimized learning rates to achieve precise fine-tuning. The model implements special token handling with no-repeat n-gram size of 3 and a repetition penalty of 3.5 to ensure diverse and natural summaries.

  • Optimized for long-form technical and narrative content
  • Implements SparkNotes-style explanatory summaries
  • Supports batch processing of extensive documents
  • Features customizable generation parameters

Core Capabilities

  • Processes documents up to 16,384 tokens in length
  • Generates abstractive summaries with explanatory elements
  • Handles technical, academic, and narrative content effectively
  • Achieves 33.45 ROUGE-1 score on benchmark tests

Frequently Asked Questions

Q: What makes this model unique?

This model's unique strength lies in its ability to handle extremely long documents while producing coherent, explanatory summaries. It's specifically optimized for book-length and technical content, making it ideal for educational and research applications.

Q: What are the recommended use cases?

The model is best suited for summarizing long narratives, academic papers, textbooks, and technical documents. It's particularly effective when you need to maintain important context while significantly reducing content length.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026