Give us the facts: Enhancing large language models with knowledge graphs for fact-aware language modeling

L Yang, H Chen, Z Li, X Ding… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Recently, ChatGPT, a representative large language model (LLM), has gained considerable
attention. Due to their powerful emergent abilities, recent LLMs are considered as a possible …

[HTML][HTML] Pre-trained language models and their applications

H Wang, J Li, H Wu, E Hovy, Y Sun - Engineering, 2023 - Elsevier
Pre-trained language models have achieved striking success in natural language
processing (NLP), leading to a paradigm shift from supervised learning to pre-training …

A survey of knowledge enhanced pre-trained language models

L Hu, Z Liu, Z Zhao, L Hou, L Nie… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Pre-trained Language Models (PLMs) which are trained on large text corpus via self-
supervised learning method, have yielded promising performance on various tasks in …

Recent advances in natural language processing via large pre-trained language models: A survey

B Min, H Ross, E Sulem, APB Veyseh… - ACM Computing …, 2023 - dl.acm.org
Large, pre-trained language models (PLMs) such as BERT and GPT have drastically
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …

Raptor: Recursive abstractive processing for tree-organized retrieval

P Sarthi, S Abdullah, A Tuli, S Khanna… - The Twelfth …, 2024 - openreview.net
Retrieval-augmented language models can better adapt to changes in world state and
incorporate long-tail knowledge. However, most existing methods retrieve only short …

Learning how to ask: Querying LMs with mixtures of soft prompts

G Qin, J Eisner - arxiv preprint arxiv:2104.06599, 2021 - arxiv.org
Natural-language prompts have recently been used to coax pretrained language models
into performing other AI tasks, using a fill-in-the-blank paradigm (Petroni et al., 2019) or a …

It's not just size that matters: Small language models are also few-shot learners

T Schick, H Schütze - arxiv preprint arxiv:2009.07118, 2020 - arxiv.org
When scaled to hundreds of billions of parameters, pretrained language models such as
GPT-3 (Brown et al., 2020) achieve remarkable few-shot performance. However, enormous …

Large pre-trained language models contain human-like biases of what is right and wrong to do

P Schramowski, C Turan, N Andersen… - Nature Machine …, 2022 - nature.com
Artificial writing is permeating our lives due to recent advances in large-scale, transformer-
based language models (LMs) such as BERT, GPT-2 and GPT-3. Using them as pre-trained …

Leveraging passage retrieval with generative models for open domain question answering

G Izacard, E Grave - arxiv preprint arxiv:2007.01282, 2020 - arxiv.org
Generative models for open domain question answering have proven to be competitive,
without resorting to external knowledge. While promising, this approach requires to use …

[HTML][HTML] What disease does this patient have? a large-scale open domain question answering dataset from medical exams

D **, E Pan, N Oufattole, WH Weng, H Fang… - Applied Sciences, 2021 - mdpi.com
Open domain question answering (OpenQA) tasks have been recently attracting more and
more attention from the natural language processing (NLP) community. In this work, we …