ms_bert

Maintained By
NLP4H

MS-BERT

PropertyValue
AuthorNLP4H
Base ModelBLUE-BERT base
Model URLHugging Face

What is ms_bert?

MS-BERT is a specialized BERT model pre-trained on neurological examination notes from Multiple Sclerosis (MS) patients. The model was developed using approximately 75,000 clinical notes from about 5,000 patients at St. Michael's Hospital in Toronto, containing over 35.7 million words. This domain-specific adaptation makes it particularly valuable for MS-related clinical text analysis.

Implementation Details

The model builds upon the BLUE-BERT base architecture and was further pre-trained using masked language modeling. The training data underwent careful preprocessing to ensure patient privacy, with sensitive information being systematically replaced with standardized tokens while maintaining semantic relevance.

  • Pre-trained on 75,000 clinical notes (35.7M words)
  • Data collected from 2015 to 2019
  • Gender distribution: 72% female, 28% male (reflecting MS prevalence)
  • Comprehensive de-identification process with semantic token replacement

Core Capabilities

  • Analysis of neurological examination notes
  • Understanding MS-specific clinical terminology
  • Processing de-identified medical text
  • Handling patient condition and progress information

Frequently Asked Questions

Q: What makes this model unique?

MS-BERT is specifically designed for Multiple Sclerosis clinical text analysis, trained on a large corpus of real-world neurological examination notes. Its specialized training makes it particularly effective for understanding MS-related medical documentation.

Q: What are the recommended use cases?

The model is designed for research purposes in processing MS-related clinical notes. However, as stated in the disclaimer, it should not be used for direct diagnostic purposes or medical decision-making without professional clinical oversight.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.