A survey of knowledge enhanced pre-trained language models
Pre-trained Language Models (PLMs) which are trained on large text corpus via self-
supervised learning method, have yielded promising performance on various tasks in …
supervised learning method, have yielded promising performance on various tasks in …
Lasuie: Unifying information extraction with latent adaptive structure-aware generative language model
Universally modeling all typical information extraction tasks (UIE) with one generative
language model (GLM) has revealed great potential by the latest study, where various IE …
language model (GLM) has revealed great potential by the latest study, where various IE …
Seeking patterns, not just memorizing procedures: Contrastive learning for solving math word problems
Math Word Problem (MWP) solving needs to discover the quantitative relationships over
natural language narratives. Recent work shows that existing models memorize procedures …
natural language narratives. Recent work shows that existing models memorize procedures …
Semantic representation for dialogue modeling
Although neural models have achieved competitive results in dialogue systems, they have
shown limited ability in representing core semantics, such as ignoring important entities. To …
shown limited ability in representing core semantics, such as ignoring important entities. To …
Structural guidance for transformer language models
Transformer-based language models pre-trained on large amounts of text data have proven
remarkably successful in learning generic transferable linguistic representations. Here we …
remarkably successful in learning generic transferable linguistic representations. Here we …
Cross-modal attention congruence regularization for vision-language relation alignment
Despite recent progress towards scaling up multimodal vision-language models, these
models are still known to struggle on compositional generalization benchmarks such as …
models are still known to struggle on compositional generalization benchmarks such as …
Emotion classification in texts over graph neural networks: Semantic representation is better than syntactic
Social media is a widely used platform that provides a huge amount of user-generated
content that can be processed to extract information about users' emotions. This has …
content that can be processed to extract information about users' emotions. This has …
Predicate-argument based bi-encoder for paraphrase identification
Paraphrase identification involves identifying whether a pair of sentences express the same
or similar meanings. While cross-encoders have achieved high performances across …
or similar meanings. While cross-encoders have achieved high performances across …
Type-driven multi-turn corrections for grammatical error correction
Grammatical Error Correction (GEC) aims to automatically detect and correct grammatical
errors. In this aspect, dominant models are trained by one-iteration learning while …
errors. In this aspect, dominant models are trained by one-iteration learning while …
Rethinking positional encoding in tree transformer for code representation
Transformers are now widely used in code representation, and several recent works further
develop tree Transformers to capture the syntactic structure in source code. Specifically …
develop tree Transformers to capture the syntactic structure in source code. Specifically …