bert-tiny-5-finetuned-squadv2

bert-tiny-5-finetuned-squadv2

mrm8488

BERT-Tiny model fine-tuned on SQuAD v2.0 for question-answering tasks. Achieves 57.12% EM and 60.86% F1 score. Compact 24.34MB size ideal for resource-constrained environments.

PropertyValue
Model Size24.34 MB
Research PaperWell-Read Students Learn Better
Performance (EM/F1)57.12% / 60.86%
Training DatasetSQuAD 2.0 (130k samples)

What is bert-tiny-5-finetuned-squadv2?

This is a compressed version of BERT, specifically designed for environments with limited computational resources. The model has been fine-tuned on the SQuAD v2.0 dataset for question-answering tasks, achieving impressive performance metrics while maintaining a compact size of just 24.34MB.

Implementation Details

The model was trained on a Tesla P100 GPU with 25GB RAM, using the SQuAD 2.0 dataset which combines 100,000 answerable questions with 50,000 unanswerable questions. It demonstrates significant improvements over its predecessor, achieving 57.12% Exact Match (EM) and 60.86% F1 score.

  • Optimized for resource-constrained environments
  • Trained using WordPiece masking
  • Capable of handling unanswerable questions
  • Easy integration with HuggingFace transformers pipeline

Core Capabilities

  • Question answering on given context
  • Ability to determine when no answer is supported by the context
  • Efficient inference with minimal computational requirements
  • Support for English language processing

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its excellent balance between size and performance. At just 24.34MB, it achieves significantly better results (57.12% EM, 60.86% F1) compared to its predecessor while maintaining a compact form factor suitable for deployment in resource-constrained environments.

Q: What are the recommended use cases?

The model is ideal for applications requiring question-answering capabilities in resource-limited settings, such as mobile devices or edge computing scenarios. It's particularly effective when used in knowledge distillation contexts, where it can benefit from larger teacher models.

Related Models

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026