Abstract
Background: With increasing rates of polypharmacy, the vigilant surveillance of clinical
drug toxicity has emerged as an important concern. Named Entity Recognition (NER) stands as
an indispensable undertaking, essential for the extraction of valuable insights regarding drug safety
from the biomedical literature. In recent years, significant advancements have been achieved in the
deep learning models on NER tasks. Nonetheless, the effectiveness of these NER techniques relies on
the availability of substantial volumes of annotated data, which is labor-intensive and inefficient.
Methods: This study introduces a novel approach that diverges from the conventional reliance on
manually annotated data. It employs a transformer-based technique known as Positive-Unlabeled
Learning (PULearning), which incorporates adaptive learning and is applied to the clinical cancer
drug toxicity corpus. To improve the precision of prediction, we employ relative position embeddings
within the transformer encoder. Additionally, we formulate a composite loss function that integrates
two Kullback-Leibler (KL) regularizers to align with PULearning assumptions. The outcomes
demonstrate that our approach attains the targeted performance for NER tasks, solely relying
on unlabeled data and named entity dictionaries.
Conclusion: Our model achieves an overall NER performance with an F1 of 0.819. Specifically, it
attains F1 of 0.841, 0.801 and 0.815 for DRUG, CANCER, and TOXI entities, respectively. A
comprehensive analysis of the results validates the effectiveness of our approach in comparison to
existing PULearning methods on biomedical NER tasks. Additionally, a visualization of the associations
among three identified entities is provided, offering a valuable reference for querying their
interrelationships.
Keywords:
KL regularizers, clinical drug toxicity, named entity recognition (NER), positive-unlabeled learning (PULearning), adaptive sampling, cancer drug.
Graphical Abstract
[14]
XU DJ. The application of text mining in social science research: Present situation, problems and prospects. Sci Soc 2015; 5(3): 75-89.
[19]
Devlin J, Chang MW, Lee K, Toutanova K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018; 2018: 04805.
[22]
Khan MR, Ziyadi M. Mt-bioner: Multi-task learning for biomedical named entity recognition using deep bidirectional transformers. arXiv 2020; 2020: 08904.
[24]
Liang C, Yu Y, Jiang H. Bond: Bert-assisted open-domain named entity recognition with distant supervision. Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining. 1054-64.
[26]
Yang Y, Chen W, Li Z. Distantly supervised NER with partial annotation learning and reinforcement learning. Proceedings of the 27th International Conference on Computational Linguistics. 2159-69.
[47]
Stout NL, Baima J, Swisher AK, Winters-Stone KM, Welsh J. A systematic review of exercise systematic reviews in the cancer literature (2005-2017). PM R 2017; 9(S2): 347-8.
[49]
Silakari O, Singh PK. Concepts and experimental protocols of modelling and informatics in drug design. Academic Press 2020.
[51]
Zhou K, Li Y, Li Q. Distantly supervised named entity recognition via confidence-based multi-class positive and unlabeled learning. arXiv 2020; 2020: 09589.
[52]
Du Plessis MC, Niu G, Sugiyama M. Analysis of learning from positive and unlabeled data. Adv Neural Inf Process Syst 2014; 27(27): 703-11.
[56]
Vaswani A, Shazeer N, Parmar N. Attention is all you need. Adv Neural Inf Process Syst 2017; 30: 5998-6008.
[57]
Zhang S, Fan R, Liu Y, Chen S, Liu Q, Zeng W. Applications of transformer-based language models in bioinformatics: A survey. Bioinform Adv 2023; 3(1): vbad001.
[62]
Yan H, Deng B, Li X. TENER: Adapting transformer encoder for named entity recognition. arXiv 2019; 2019: 1911-04474.
[64]
Liu X, Yu HF, Dhillon I. Learning to encode position for transformer with continuous dynamical model. Int Conf Mach LearnPMLR 2020; 6327-35.
[65]
Press O, Smith NA, Lewis M. Shortformer: Better language modeling using shorter inputs. arXiv 2020; 2020-15932.
[71]
Xie J, Girshick R, Farhadi A. Unsupervised deep embedding for clustering analysis. Int Conf Mach Learn PMLR 2016; 478-87.