| ### [Domain Sensitive Fine-tuning](https://github.com/mykelismyname/MSLM): | |
| Model is built by training BERT on a biomedical dataset BC2GM using an approach that learns mask specific losses. | |
| More details in paper below, | |
| #### Citation | |
| ``` | |
| @article{abaho2024improving, | |
| title={Improving Pre-trained Language Model Sensitivity via Mask Specific losses: A case study on Biomedical NER}, | |
| author={Abaho, Micheal and Bollegala, Danushka and Leeming, Gary and Joyce, Dan and Buchan, Iain E}, | |
| journal={arXiv preprint arXiv:2403.18025}, | |
| year={2024} | |
| } | |