A survey on legal judgment prediction: Datasets, metrics, models and challenges
J Cui, X Shen, S Wen - IEEE Access, 2023 - ieeexplore.ieee.org
Legal judgment prediction (LJP) applies Natural Language Processing (NLP) techniques to
predict judgment results based on fact descriptions automatically. The present work …
predict judgment results based on fact descriptions automatically. The present work …
Ast-trans: Code summarization with efficient tree-structured attention
Code summarization aims to generate brief natural language descriptions for source codes.
The state-of-the-art approaches follow a transformer-based encoder-decoder architecture …
The state-of-the-art approaches follow a transformer-based encoder-decoder architecture …
Math word problem generation with mathematical consistency and problem context constraints
We study the problem of generating arithmetic math word problems (MWPs) given a math
equation that specifies the mathematical computation and a context that specifies the …
equation that specifies the mathematical computation and a context that specifies the …
Diversifying dialogue generation with non-conversational text
Neural network-based sequence-to-sequence (seq2seq) models strongly suffer from the low-
diversity problem when it comes to open-domain dialogue generation. As bland and generic …
diversity problem when it comes to open-domain dialogue generation. As bland and generic …
Neural data-to-text generation via jointly learning the segmentation and correspondence
The neural attention model has achieved great success in data-to-text generation tasks.
Though usually excelling at producing fluent text, it suffers from the problem of information …
Though usually excelling at producing fluent text, it suffers from the problem of information …
Welm: A well-read pre-trained language model for chinese
Large Language Models pre-trained with self-supervised learning have demonstrated
impressive zero-shot generalization capabilities on a wide spectrum of tasks. In this work …
impressive zero-shot generalization capabilities on a wide spectrum of tasks. In this work …
Towards faithfulness in open domain table-to-text generation from an entity-centric view
In open domain table-to-text generation, we notice the unfaithful generation usually contains
hallucinated entities which can not be aligned to any input table record. We thus try to …
hallucinated entities which can not be aligned to any input table record. We thus try to …
Moviechats: Chat like humans in a closed domain
Being able to perform in-depth chat with humans in a closed domain is a precondition before
an open-domain chatbot can be ever claimed. In this work, we take a close look at the movie …
an open-domain chatbot can be ever claimed. In this work, we take a close look at the movie …
Ed2lm: Encoder-decoder to language model for faster document re-ranking inference
State-of-the-art neural models typically encode document-query pairs using cross-attention
for re-ranking. To this end, models generally utilize an encoder-only (like BERT) paradigm or …
for re-ranking. To this end, models generally utilize an encoder-only (like BERT) paradigm or …
DYPLOC: Dynamic planning of content using mixed language models for text generation
We study the task of long-form opinion text generation, which faces at least two distinct
challenges. First, existing neural generation models fall short of coherence, thus requiring …
challenges. First, existing neural generation models fall short of coherence, thus requiring …