xlm-roberta-ner-japanese

Maintained By
tsmatz

XLM-RoBERTa Japanese NER Model

PropertyValue
Parameter Count277M
LicenseMIT
Model TypeToken Classification
FrameworkPyTorch with Transformers
F1 Score0.9864

What is xlm-roberta-ner-japanese?

The xlm-roberta-ner-japanese is a specialized Named Entity Recognition (NER) model fine-tuned on Japanese Wikipedia articles. Built upon the xlm-roberta-base architecture, this model excels at identifying and classifying named entities in Japanese text, including persons, organizations, locations, and more.

Implementation Details

The model was fine-tuned using the Stockmark Inc. dataset with carefully selected hyperparameters including a learning rate of 5e-05 and trained over 5 epochs. It utilizes the PyTorch framework and Transformers library, achieving impressive validation metrics with a final F1 score of 0.9864.

  • Trained with Adam optimizer (betas=0.9,0.999)
  • Batch size of 12 for both training and evaluation
  • Linear learning rate scheduler
  • Supports 9 distinct entity categories

Core Capabilities

  • Person (PER) identification
  • Organization detection (ORG, ORG-P, ORG-O)
  • Location (LOC) recognition
  • Institution/facility (INS) classification
  • Product (PRD) and Event (EVT) detection

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its specialized Japanese NER capabilities, trained on high-quality Wikipedia data with a comprehensive entity classification system. Its high F1 score of 0.9864 demonstrates exceptional accuracy in entity recognition tasks.

Q: What are the recommended use cases?

The model is ideal for Japanese text analysis tasks requiring entity extraction, such as information retrieval, content categorization, and automated text processing. It can be easily implemented using the Transformers pipeline for token classification.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.