DeepStruct: Pretraining of language models for structure prediction

C Wang, X Liu, Z Chen, H Hong, J Tang… - arxiv preprint arxiv …, 2022 - arxiv.org
We introduce a method for improving the structural understanding abilities of language
models. Unlike previous approaches that finetune the models with task-specific …

Universal information extraction as unified semantic matching

J Lou, Y Lu, D Dai, W Jia, H Lin, X Han… - Proceedings of the AAAI …, 2023 - ojs.aaai.org
The challenge of information extraction (IE) lies in the diversity of label schemas and the
heterogeneity of structures. Traditional methods require task-specific model design and rely …

Codekgc: Code language model for generative knowledge graph construction

Z Bi, J Chen, Y Jiang, F **ong, W Guo, H Chen… - ACM Transactions on …, 2024 - dl.acm.org
Current generative knowledge graph construction approaches usually fail to capture
structural knowledge by simply flattening natural language into serialized texts or a …

Knowledge mining: A cross-disciplinary survey

Y Rui, VIS Carmona, M Pourvali, Y **ng, WW Yi… - Machine Intelligence …, 2022 - Springer
Abstract Knowledge mining is a widely active research area across disciplines such as
natural language processing (NLP), data mining (DM), and machine learning (ML). The …

[PDF][PDF] Bertnet: Harvesting knowledge graphs from pretrained language models

S Hao, B Tan, K Tang, H Zhang… - arxiv preprint arxiv …, 2022 - researchgate.net
Symbolic knowledge graphs (KGs) have been constructed either by expensive human
crowdsourcing or with complex text mining pipelines. The emerging large pretrained …

Glitter or gold? Deriving structured insights from sustainability reports via large language models

M Bronzini, C Nicolini, B Lepri, A Passerini… - EPJ Data …, 2024 - epjds.epj.org
Over the last decade, several regulatory bodies have started requiring the disclosure of non-
financial information from publicly listed companies, in light of the investors' increasing …

Generative prompt tuning for relation classification

J Han, S Zhao, B Cheng, S Ma, W Lu - arxiv preprint arxiv:2210.12435, 2022 - arxiv.org
Using prompts to explore the knowledge contained within pre-trained language models for
downstream tasks has now become an active topic. Current prompt tuning methods mostly …

[PDF][PDF] Open information extraction from 2007 to 2022–a survey

P Liu, W Gao, W Dong, S Huang… - arxiv preprint arxiv …, 2022 - researchgate.net
Open information extraction is an important NLP task that targets extracting structured
information from unstructured text without limitations on the relation type or the domain of the …

Evaluation and analysis of large language models for clinical text augmentation and generation

A Latif, J Kim - IEEE Access, 2024 - ieeexplore.ieee.org
A major challenge in deep learning (DL) model training is data scarcity. Data scarcity is
commonly found in specific domains, such as clinical or low-resource languages, that are …

Benchmarking language models for code syntax understanding

D Shen, X Chen, C Wang, K Sen, D Song - arxiv preprint arxiv …, 2022 - arxiv.org
Pre-trained language models have demonstrated impressive performance in both natural
language processing and program understanding, which represent the input as a token …