Document-level machine translation with large language models

L Wang, C Lyu, T Ji, Z Zhang, D Yu, S Shi… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) such as ChatGPT can produce coherent, cohesive, relevant,
and fluent answers for various natural language processing (NLP) tasks. Taking document …

Incremental transformer structure enhanced image inpainting with masking positional encoding

Q Dong, C Cao, Y Fu - … of the IEEE/CVF conference on …, 2022 - openaccess.thecvf.com
Image inpainting has made significant advances in recent years. However, it is still
challenging to recover corrupted images with both vivid textures and reasonable structures …

Break the sequential dependency of llm inference using lookahead decoding

Y Fu, P Bailis, I Stoica, H Zhang - arxiv preprint arxiv:2402.02057, 2024 - arxiv.org
Autoregressive decoding of large language models (LLMs) is memory bandwidth bounded,
resulting in high latency and significant wastes of the parallel processing power of modern …

On the effectiveness of adapter-based tuning for pretrained language model adaptation

R He, L Liu, H Ye, Q Tan, B Ding, L Cheng… - arxiv preprint arxiv …, 2021 - arxiv.org
Adapter-based tuning has recently arisen as an alternative to fine-tuning. It works by adding
light-weight adapter modules to a pretrained language model (PrLM) and only updating the …

A survey on non-autoregressive generation for neural machine translation and beyond

Y **ao, L Wu, J Guo, J Li, M Zhang… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Non-autoregressive (NAR) generation, which is first proposed in neural machine translation
(NMT) to speed up inference, has attracted much attention in both machine learning and …

Diffusion language models are versatile protein learners

X Wang, Z Zheng, F Ye, D Xue, S Huang… - arxiv preprint arxiv …, 2024 - arxiv.org
This paper introduces diffusion protein language model (DPLM), a versatile protein
language model that demonstrates strong generative and predictive capabilities for protein …

UFC-BERT: Unifying multi-modal controls for conditional image synthesis

Z Zhang, J Ma, C Zhou, R Men, Z Li… - Advances in …, 2021 - proceedings.neurips.cc
Conditional image synthesis aims to create an image according to some multi-modal
guidance in the forms of textual descriptions, reference images, and image blocks to …

How to design translation prompts for ChatGPT: An empirical study

Y Gao, R Wang, F Hou - Proceedings of the 6th ACM International …, 2024 - dl.acm.org
ChatGPT, a chatbot based on the GPT models, has demonstrated surprising abilities in
natural language understanding and generation tasks. Given that machine translation …

Fast nearest neighbor machine translation

Y Meng, X Li, X Zheng, F Wu, X Sun, T Zhang… - arxiv preprint arxiv …, 2021 - arxiv.org
Though nearest neighbor Machine Translation ($ k $ NN-MT)\citep {khandelwal2020nearest
} has proved to introduce significant performance boosts over standard neural MT systems, it …

MSP: Multi-stage prompting for making pre-trained language models better translators

Z Tan, X Zhang, S Wang, Y Liu - arxiv preprint arxiv:2110.06609, 2021 - arxiv.org
Prompting has recently been shown as a promising approach for applying pre-trained
language models to perform downstream tasks. We present Multi-Stage Prompting (MSP), a …