[HTML][HTML] ChatGPT in healthcare: a taxonomy and systematic review

J Li, A Dada, B Puladi, J Kleesiek, J Egger - Computer Methods and …, 2024 - Elsevier
The recent release of ChatGPT, a chat bot research project/product of natural language
processing (NLP) by OpenAI, stirs up a sensation among both the general public and …

Deep learning--based text classification: a comprehensive review

S Minaee, N Kalchbrenner, E Cambria… - ACM computing …, 2021 - dl.acm.org
Deep learning--based models have surpassed classical machine learning--based
approaches in various text classification tasks, including sentiment analysis, news …

Raft: Adapting language model to domain specific rag

T Zhang, SG Patil, N Jain, S Shen… - First Conference on …, 2024 - openreview.net
Pretraining Large Language Models (LLMs) on large corpora of textual data is now a
standard paradigm. When using these LLMs for many downstream applications, it is …

Data selection for language models via importance resampling

SM **e, S Santurkar, T Ma… - Advances in Neural …, 2023 - proceedings.neurips.cc
Selecting a suitable pretraining dataset is crucial for both general-domain (eg, GPT-3) and
domain-specific (eg, Codex) language models (LMs). We formalize this problem as selecting …

Increasing diversity while maintaining accuracy: Text data generation with large language models and human interventions

JJY Chung, E Kamar, S Amershi - arxiv preprint arxiv:2306.04140, 2023 - arxiv.org
Large language models (LLMs) can be used to generate text data for training and evaluating
other models. However, creating high-quality datasets with LLMs can be challenging. In this …

Adapting Large Language Models to Domains via Reading Comprehension

D Cheng, S Huang, F Wei - arxiv preprint arxiv:2309.09530, 2023 - arxiv.org
We explore how continued pre-training on domain-specific corpora influences large
language models, revealing that training on the raw corpora endows the model with domain …

Don't stop pretraining: Adapt language models to domains and tasks

S Gururangan, A Marasović, S Swayamdipta… - arxiv preprint arxiv …, 2020 - arxiv.org
Language models pretrained on text from a wide variety of sources form the foundation of
today's NLP. In light of the success of these broad-coverage models, we investigate whether …

Active learning by acquiring contrastive examples

K Margatina, G Vernikos, L Barrault… - arxiv preprint arxiv …, 2021 - arxiv.org
Common acquisition functions for active learning use either uncertainty or diversity
sampling, aiming to select difficult and diverse data points from the pool of unlabeled data …

AI vs. Human--differentiation analysis of scientific content generation

Y Ma, J Liu, F Yi, Q Cheng, Y Huang, W Lu… - arxiv preprint arxiv …, 2023 - arxiv.org
Recent neural language models have taken a significant step forward in producing
remarkably controllable, fluent, and grammatical text. Although studies have found that AI …

Cold-start active learning through self-supervised language modeling

M Yuan, HT Lin, J Boyd-Graber - arxiv preprint arxiv:2010.09535, 2020 - arxiv.org
Active learning strives to reduce annotation costs by choosing the most critical examples to
label. Typically, the active learning strategy is contingent on the classification model. For …