mBERT-base-Biomedical-NER

Maintained By
StivenLancheros

mBERT-base-Biomedical-NER

PropertyValue
LicenseApache 2.0
FrameworkPyTorch, Transformers
Training DataCRAFT+BC4CHEMD+BioNLP09
PerformanceF1: 0.9831, Accuracy: 0.9799

What is mBERT-base-Biomedical-NER?

This model is a specialized fine-tuned version of bert-base-multilingual-cased, designed specifically for biomedical named entity recognition tasks. It represents a significant advancement in multilingual biomedical text analysis, achieving exceptional performance metrics with 98.31% F1 score.

Implementation Details

The model was trained using carefully selected hyperparameters, including a learning rate of 3e-05 and Adam optimizer with betas=(0.9,0.999). The training process spanned 4 epochs with a batch size of 8, utilizing a linear learning rate scheduler.

  • Precision: 0.9830
  • Recall: 0.9832
  • Training Loss: 0.1027
  • Validation Performance: 0.9799 accuracy

Core Capabilities

  • Multilingual biomedical named entity recognition
  • High-precision entity extraction
  • Cross-lingual biomedical text analysis
  • Support for multiple biomedical entity types

Frequently Asked Questions

Q: What makes this model unique?

This model combines multilingual capabilities with specialized biomedical NER training, achieving state-of-the-art performance on multiple biomedical datasets. Its high F1 score of 0.9831 makes it particularly reliable for production environments.

Q: What are the recommended use cases?

The model is ideal for biomedical text analysis, research document processing, and multilingual medical entity extraction. However, as noted by the author, it's currently in testing phase and not yet recommended for professional use.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.