Pre-trained language models in biomedical domain: A systematic survey

B Wang, Q **e, J Pei, Z Chen, P Tiwari, Z Li… - ACM Computing …, 2023 - dl.acm.org
Pre-trained language models (PLMs) have been the de facto paradigm for most natural
language processing tasks. This also benefits the biomedical domain: researchers from …

Neural natural language processing for unstructured data in electronic health records: a review

I Li, J Pan, J Goldwasser, N Verma, WP Wong… - Computer Science …, 2022 - Elsevier
Electronic health records (EHRs), digital collections of patient healthcare events and
observations, are ubiquitous in medicine and critical to healthcare delivery, operations, and …

Publicly available clinical BERT embeddings

E Alsentzer, JR Murphy, W Boag, WH Weng… - arxiv preprint arxiv …, 2019 - arxiv.org
Contextual word embedding models such as ELMo (Peters et al., 2018) and BERT (Devlin et
al., 2018) have dramatically improved performance for many natural language processing …

BioBERT: a pre-trained biomedical language representation model for biomedical text mining

J Lee, W Yoon, S Kim, D Kim, S Kim, CH So… - …, 2020 - academic.oup.com
Motivation Biomedical text mining is becoming increasingly important as the number of
biomedical documents rapidly grows. With the progress in natural language processing …

Pretrained language models for biomedical and clinical tasks: understanding and extending the state-of-the-art

P Lewis, M Ott, J Du, V Stoyanov - Proceedings of the 3rd clinical …, 2020 - aclanthology.org
A large array of pretrained models are available to the biomedical NLP (BioNLP) community.
Finding the best model for a particular task can be difficult and time-consuming. For many …

Enhancing clinical concept extraction with contextual embeddings

Y Si, J Wang, H Xu, K Roberts - Journal of the American Medical …, 2019 - academic.oup.com
Objective Neural network–based representations (“embeddings”) have dramatically
advanced natural language processing (NLP) tasks, including clinical NLP tasks such as …

Scifive: a text-to-text transformer model for biomedical literature

LN Phan, JT Anibal, H Tran, S Chanana… - arxiv preprint arxiv …, 2021 - arxiv.org
In this report, we introduce SciFive, a domain-specific T5 model that has been pre-trained on
large biomedical corpora. Our model outperforms the current SOTA methods (ie BERT …

Deep learning in clinical natural language processing: a methodical review

S Wu, K Roberts, S Datta, J Du, Z Ji, Y Si… - Journal of the …, 2020 - academic.oup.com
Objective This article methodically reviews the literature on deep learning (DL) for natural
language processing (NLP) in the clinical domain, providing quantitative analysis to answer …

[HTML][HTML] A survey of word embeddings for clinical text

FK Khattak, S Jeblee, C Pou-Prom, M Abdalla… - Journal of Biomedical …, 2019 - Elsevier
Representing words as numerical vectors based on the contexts in which they appear has
become the de facto method of analyzing text with machine learning. In this paper, we …

Umlsbert: Clinical domain knowledge augmentation of contextual embeddings using the unified medical language system metathesaurus

G Michalopoulos, Y Wang, H Kaka, H Chen… - arxiv preprint arxiv …, 2020 - arxiv.org
Contextual word embedding models, such as BioBERT and Bio_ClinicalBERT, have
achieved state-of-the-art results in biomedical natural language processing tasks by …