Towards Non-invasive and Personalized Management of Breast Cancer Patients from Multiparametric MRI via A Large Mixture-of-Modality-Experts Model
MOME conducts multimodel fusion and classification based on multi-sequence 3D medical data, e.g., multiparametric breast MRI.
- Developed by: Luyang Luo
- Model type: Transformer (based on BEiT3)
- License: MIT
- Finetuned from model : BEiT-3
- Repository: https://github.com/LLYXC/MOME
- Paper: Towards Non-invasive and Personalized Management of Breast Cancer Patients from Multiparametric MRI via A Large Mixture-of-Modality-Experts Model
- Requirement/dependencies: Please see the requirement of BEiT-3.
- Installation: The installation will take a few seconds to minutes.
git clone https://github.com/LLYXC/MOME.git
mkdir log
- Sample Data
- Training and Testing: The training and testing commands are provided in scripts. Make a directory for the dataset, and use the csv file provided above to load the dataset.
- Expected output: The model will generate probabilities for each sample, and the expected run time on an NVIDIA GeForce RTX 3090 GPU will be less than one minute.
@article{luo2024towards,
title={Towards Non-invasive and Personalized Management of Breast Cancer Patients from Multiparametric MRIvia A Large Mixture-of-Modality-Experts Model},
author={Luo, Luyang and Wu, Mingxiang and Li, Mei and Xin, Yi and Wang, Qiong and Vardhanabhuti, Varut andChu, Winnie CW and Li, Zhenhui and Zhou, Juan and Rajpurkar, Pranav and Chen, Hao},
year={2024}
}
Our work is standing on the sholders of BEiT-3 and soft MOE, please also consider cite the following works:
@inproceedings{wang2023image,
title={Image as a foreign language: Beit pretraining for vision and vision-language tasks},
author={Wang, Wenhui and Bao, Hangbo and Dong, Li and Bjorck, Johan and Peng, Zhiliang and Liu, Qiang and Aggarwal, Kriti and Mohammed, Owais Khan and Singhal, Saksham and Som, Subhojit and others},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages={19175--19186},
year={2023}
}
@inproceedings{puigcerversparse,
title={From Sparse to Soft Mixtures of Experts},
author={Puigcerver, Joan and Ruiz, Carlos Riquelme and Mustafa, Basil and Houlsby, Neil},
booktitle={The Twelfth International Conference on Learning Representations}
}