A survey on legal judgment prediction: Datasets, metrics, models and challenges

J Cui, X Shen, S Wen - IEEE Access, 2023 - ieeexplore.ieee.org
Legal judgment prediction (LJP) applies Natural Language Processing (NLP) techniques to
predict judgment results based on fact descriptions automatically. The present work …

Ast-trans: Code summarization with efficient tree-structured attention

Z Tang, X Shen, C Li, J Ge, L Huang, Z Zhu… - Proceedings of the 44th …, 2022 - dl.acm.org
Code summarization aims to generate brief natural language descriptions for source codes.
The state-of-the-art approaches follow a transformer-based encoder-decoder architecture …

Math word problem generation with mathematical consistency and problem context constraints

Z Wang, AS Lan, RG Baraniuk - arxiv preprint arxiv:2109.04546, 2021 - arxiv.org
We study the problem of generating arithmetic math word problems (MWPs) given a math
equation that specifies the mathematical computation and a context that specifies the …

Diversifying dialogue generation with non-conversational text

H Su, X Shen, S Zhao, X Zhou, P Hu, R Zhong… - arxiv preprint arxiv …, 2020 - arxiv.org
Neural network-based sequence-to-sequence (seq2seq) models strongly suffer from the low-
diversity problem when it comes to open-domain dialogue generation. As bland and generic …

Neural data-to-text generation via jointly learning the segmentation and correspondence

X Shen, E Chang, H Su, J Zhou, D Klakow - arxiv preprint arxiv …, 2020 - arxiv.org
The neural attention model has achieved great success in data-to-text generation tasks.
Though usually excelling at producing fluent text, it suffers from the problem of information …

Welm: A well-read pre-trained language model for chinese

H Su, X Zhou, H Yu, X Shen, Y Chen, Z Zhu… - arxiv preprint arxiv …, 2022 - arxiv.org
Large Language Models pre-trained with self-supervised learning have demonstrated
impressive zero-shot generalization capabilities on a wide spectrum of tasks. In this work …

Towards faithfulness in open domain table-to-text generation from an entity-centric view

T Liu, X Zheng, B Chang, Z Sui - … of the AAAI Conference on Artificial …, 2021 - ojs.aaai.org
In open domain table-to-text generation, we notice the unfaithful generation usually contains
hallucinated entities which can not be aligned to any input table record. We thus try to …

Moviechats: Chat like humans in a closed domain

H Su, X Shen, Z **ao, Z Zhang, E Chang… - Proceedings of the …, 2020 - aclanthology.org
Being able to perform in-depth chat with humans in a closed domain is a precondition before
an open-domain chatbot can be ever claimed. In this work, we take a close look at the movie …

Ed2lm: Encoder-decoder to language model for faster document re-ranking inference

K Hui, H Zhuang, T Chen, Z Qin, J Lu, D Bahri… - arxiv preprint arxiv …, 2022 - arxiv.org
State-of-the-art neural models typically encode document-query pairs using cross-attention
for re-ranking. To this end, models generally utilize an encoder-only (like BERT) paradigm or …

DYPLOC: Dynamic planning of content using mixed language models for text generation

X Hua, A Sreevatsa, L Wang - arxiv preprint arxiv:2106.00791, 2021 - arxiv.org
We study the task of long-form opinion text generation, which faces at least two distinct
challenges. First, existing neural generation models fall short of coherence, thus requiring …