A comprehensive survey on automatic knowledge graph construction

L Zhong, J Wu, Q Li, H Peng, X Wu - ACM Computing Surveys, 2023 - dl.acm.org
Automatic knowledge graph construction aims at manufacturing structured human
knowledge. To this end, much effort has historically been spent extracting informative fact …

Unifying large language models and knowledge graphs: A roadmap

S Pan, L Luo, Y Wang, C Chen… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Large language models (LLMs), such as ChatGPT and GPT4, are making new waves in the
field of natural language processing and artificial intelligence, due to their emergent ability …

Symbolic knowledge distillation: from general language models to commonsense models

P West, C Bhagavatula, J Hessel, JD Hwang… - arxiv preprint arxiv …, 2021 - arxiv.org
The common practice for training commonsense models has gone from-human-to-corpus-to-
machine: humans author commonsense knowledge graphs in order to train commonsense …

Open-world story generation with structured knowledge enhancement: A comprehensive survey

Y Wang, J Lin, Z Yu, W Hu, BF Karlsson - Neurocomputing, 2023 - Elsevier
Storytelling and narrative are fundamental to human experience, intertwined with our social
and cultural engagement. As such, researchers have long attempted to create systems that …

K-lite: Learning transferable visual models with external knowledge

S Shen, C Li, X Hu, Y **e, J Yang… - Advances in …, 2022 - proceedings.neurips.cc
The new generation of state-of-the-art computer vision systems are trained from natural
language supervision, ranging from simple object category names to descriptive captions …

(comet-) atomic 2020: On symbolic and neural commonsense knowledge graphs

JD Hwang, C Bhagavatula, R Le Bras, J Da… - Proceedings of the …, 2021 - ojs.aaai.org
Recent years have brought about a renewed interest in commonsense representation and
reasoning in the field of natural language understanding. The development of new …

A survey of knowledge-intensive nlp with pre-trained language models

D Yin, L Dong, H Cheng, X Liu, KW Chang… - arxiv preprint arxiv …, 2022 - arxiv.org
With the increasing of model capacity brought by pre-trained language models, there
emerges boosting needs for more knowledgeable natural language processing (NLP) …

Maven-ere: A unified large-scale dataset for event coreference, temporal, causal, and subevent relation extraction

X Wang, Y Chen, N Ding, H Peng, Z Wang… - arxiv preprint arxiv …, 2022 - arxiv.org
The diverse relationships among real-world events, including coreference, temporal, causal,
and subevent relations, are fundamental to understanding natural languages. However, two …

Improving large language models in event relation logical prediction

M Chen, Y Ma, K Song, Y Cao… - Proceedings of the 62nd …, 2024 - aclanthology.org
Event relations are crucial for narrative understanding and reasoning. Governed by nuanced
logic, event relation extraction (ERE) is a challenging task that demands thorough semantic …

Extracting cultural commonsense knowledge at scale

TP Nguyen, S Razniewski, A Varde… - Proceedings of the ACM …, 2023 - dl.acm.org
Structured knowledge is important for many AI applications. Commonsense knowledge,
which is crucial for robust human-centric AI, is covered by a small number of structured …