A survey of large language models in medicine: Progress, application, and challenge

H Zhou, F Liu, B Gu, X Zou, J Huang, J Wu, Y Li… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs), such as ChatGPT, have received substantial attention due
to their capabilities for understanding and generating human language. While there has …

[HTML][HTML] Natural language processing for innovation search–Reviewing an emerging non-human innovation intermediary

J Just - Technovation, 2024 - Elsevier
Applying artificial intelligence (AI), especially natural language processing (NLP), to harness
large amounts of information from patent databases, online communities, social media, or …

Large language models are few-shot clinical information extractors

M Agrawal, S Hegselmann, H Lang, Y Kim… - arxiv preprint arxiv …, 2022 - arxiv.org
A long-running goal of the clinical NLP community is the extraction of important variables
trapped in clinical notes. However, roadblocks have included dataset shift from the general …

The values encoded in machine learning research

A Birhane, P Kalluri, D Card, W Agnew… - Proceedings of the …, 2022 - dl.acm.org
Machine learning currently exerts an outsized influence on the world, increasingly affecting
institutional practices and impacted communities. It is therefore critical that we question …

Energy efficient semantic communication over wireless networks with rate splitting

Z Yang, M Chen, Z Zhang… - IEEE Journal on Selected …, 2023 - ieeexplore.ieee.org
In this paper, the problem of wireless resource allocation and semantic information
extraction for energy efficient semantic communications over wireless networks with rate …

Performance optimization for semantic communications: An attention-based reinforcement learning approach

Y Wang, M Chen, T Luo, W Saad… - IEEE Journal on …, 2022 - ieeexplore.ieee.org
In this paper, a semantic communication framework is proposed for textual data
transmission. In the studied model, a base station (BS) extracts the semantic information …

Specter: Document-level representation learning using citation-informed transformers

A Cohan, S Feldman, I Beltagy, D Downey… - arxiv preprint arxiv …, 2020 - arxiv.org
Representation learning is a critical ingredient for natural language processing systems.
Recent Transformer language models like BERT learn powerful textual representations, but …

The semantic scholar open data platform

R Kinney, C Anastasiades, R Authur, I Beltagy… - arxiv preprint arxiv …, 2023 - arxiv.org
The volume of scientific output is creating an urgent need for automated tools to help
scientists keep up with developments in their field. Semantic Scholar (S2) is an open data …

SciBERT: A pretrained language model for scientific text

I Beltagy, K Lo, A Cohan - arxiv preprint arxiv:1903.10676, 2019 - arxiv.org
Obtaining large-scale annotated data for NLP tasks in the scientific domain is challenging
and expensive. We release SciBERT, a pretrained language model based on BERT (Devlin …

MatSciBERT: A materials domain language model for text mining and information extraction

T Gupta, M Zaki, NMA Krishnan, Mausam - npj Computational Materials, 2022 - nature.com
A large amount of materials science knowledge is generated and stored as text published in
peer-reviewed scientific literature. While recent developments in natural language …