tner-xlm-roberta-base-ontonotes5

Maintained By
asahi417

tner-xlm-roberta-base-ontonotes5

PropertyValue
AuthorAsahi Ushio
Model TypeToken Classification
Base ArchitectureXLM-RoBERTa
PaperResearch Paper
Downloads18,395
LanguageEnglish

What is tner-xlm-roberta-base-ontonotes5?

This is a specialized Named Entity Recognition (NER) model built on the XLM-RoBERTa architecture and fine-tuned on the OntoNotes5 dataset. It features 12 attention heads and 12 hidden layers, with a vocabulary size of 250,002 tokens, making it particularly effective for multilingual token classification tasks.

Implementation Details

The model utilizes the transformer architecture with key specifications including a layer normalization epsilon of 1e-05 and is implemented using PyTorch. It's designed to work seamlessly with the tner library and can be easily integrated into existing NLP pipelines.

  • 12 attention heads for enhanced context understanding
  • 12 hidden layers for deep feature extraction
  • Comprehensive token classification capabilities
  • Support for multiple entity types including ORG, MISC, PER, and LOC

Core Capabilities

  • Named Entity Recognition across multiple languages
  • Token-level classification with 9 distinct label categories
  • Integration with popular NLP frameworks
  • Support for both direct and downstream applications

Frequently Asked Questions

Q: What makes this model unique?

This model combines the multilingual capabilities of XLM-RoBERTa with specialized NER training on OntoNotes5, making it particularly effective for cross-lingual named entity recognition tasks.

Q: What are the recommended use cases?

The model is ideal for tasks requiring named entity recognition in multilingual contexts, including organization name detection, person name identification, location recognition, and miscellaneous entity classification.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.