A survey of knowledge enhanced pre-trained language models

L Hu, Z Liu, Z Zhao, L Hou, L Nie… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Pre-trained Language Models (PLMs) which are trained on large text corpus via self-
supervised learning method, have yielded promising performance on various tasks in …

Lasuie: Unifying information extraction with latent adaptive structure-aware generative language model

H Fei, S Wu, J Li, B Li, F Li, L Qin… - Advances in …, 2022 - proceedings.neurips.cc
Universally modeling all typical information extraction tasks (UIE) with one generative
language model (GLM) has revealed great potential by the latest study, where various IE …

Seeking patterns, not just memorizing procedures: Contrastive learning for solving math word problems

Z Li, W Zhang, C Yan, Q Zhou, C Li, H Liu… - arxiv preprint arxiv …, 2021 - arxiv.org
Math Word Problem (MWP) solving needs to discover the quantitative relationships over
natural language narratives. Recent work shows that existing models memorize procedures …

Semantic representation for dialogue modeling

X Bai, Y Chen, L Song, Y Zhang - arxiv preprint arxiv:2105.10188, 2021 - arxiv.org
Although neural models have achieved competitive results in dialogue systems, they have
shown limited ability in representing core semantics, such as ignoring important entities. To …

Structural guidance for transformer language models

P Qian, T Naseem, R Levy, RF Astudillo - arxiv preprint arxiv:2108.00104, 2021 - arxiv.org
Transformer-based language models pre-trained on large amounts of text data have proven
remarkably successful in learning generic transferable linguistic representations. Here we …

Cross-modal attention congruence regularization for vision-language relation alignment

R Pandey, R Shao, PP Liang, R Salakhutdinov… - arxiv preprint arxiv …, 2022 - arxiv.org
Despite recent progress towards scaling up multimodal vision-language models, these
models are still known to struggle on compositional generalization benchmarks such as …

Emotion classification in texts over graph neural networks: Semantic representation is better than syntactic

I Ameer, N Bölücü, G Sidorov, B Can - IEEE Access, 2023 - ieeexplore.ieee.org
Social media is a widely used platform that provides a huge amount of user-generated
content that can be processed to extract information about users' emotions. This has …

Predicate-argument based bi-encoder for paraphrase identification

Q Peng, D Weir, J Weeds, Y Chai - … of the 60th Annual Meeting of …, 2022 - aclanthology.org
Paraphrase identification involves identifying whether a pair of sentences express the same
or similar meanings. While cross-encoders have achieved high performances across …

Type-driven multi-turn corrections for grammatical error correction

S Lai, Q Zhou, J Zeng, Z Li, C Li, Y Cao, J Su - arxiv preprint arxiv …, 2022 - arxiv.org
Grammatical Error Correction (GEC) aims to automatically detect and correct grammatical
errors. In this aspect, dominant models are trained by one-iteration learning while …

Rethinking positional encoding in tree transformer for code representation

H Peng, G Li, Y Zhao, Z ** - … of the 2022 Conference on Empirical …, 2022 - aclanthology.org
Transformers are now widely used in code representation, and several recent works further
develop tree Transformers to capture the syntactic structure in source code. Specifically …