Biomed V L P C X R B E R T general
microsoftIntroduction
CXR-BERT-GENERAL is a domain-specific language model tailored for chest X-ray (CXR) interpretations. It employs a specialized vocabulary, an innovative pretraining strategy, weight regularization, and text augmentations to enhance performance in radiology natural language inference, masked language model token prediction, and vision-language processing tasks like zero-shot phrase grounding and image classification.
Architecture
CXR-BERT-GENERAL is pretrained from a randomly initialized BERT model using Masked Language Modeling (MLM) on datasets including PubMed and clinical notes from MIMIC-III and MIMIC-CXR. The model's architecture enables it to be fine-tuned for various clinical research applications beyond chest radiology. CXR-BERT-specialized is an extension, further pretrained for chest X-ray domains using a multi-modal contrastive learning framework similar to CLIP.
Training
The model is trained using abstracts from PubMed and clinical notes from MIMIC datasets. The training process includes multiple phases:
- Phase I: Pretraining with MLM on combined biomedical literature and clinical data.
- Phase II: Specialization for chest X-ray data.
- Phase III: Multi-modal contrastive learning to align text and image embeddings.
Guide: Running Locally
To run CXR-BERT-GENERAL locally:
- Install Dependencies: Ensure Python and PyTorch are installed. Use
pip install transformers
for the Hugging Face Transformers library. - Download Model: Access the model via Hugging Face:
microsoft/BiomedVLP-CXR-BERT-general
. - Load Model: Use the Transformers library to load the model and tokenizer:
from transformers import AutoTokenizer, AutoModelForMaskedLM tokenizer = AutoTokenizer.from_pretrained("microsoft/BiomedVLP-CXR-BERT-general") model = AutoModelForMaskedLM.from_pretrained("microsoft/BiomedVLP-CXR-BERT-general")
- Inference: Use the model for tasks like masked language prediction.
For optimal performance, consider using cloud GPUs from providers like AWS, Google Cloud, or Azure.
License
CXR-BERT-GENERAL is released under the MIT License, allowing for wide usage and modification, provided that the license terms are met.