Statistical machine translation

A Lopez - ACM Computing Surveys (CSUR), 2008 - dl.acm.org
Statistical machine translation (SMT) treats the translation of natural language as a machine
learning problem. By examining many samples of human-produced translation, SMT …

Semantic reconstruction of continuous language from non-invasive brain recordings

J Tang, A LeBel, S Jain, AG Huth - Nature Neuroscience, 2023 - nature.com
A brain–computer interface that decodes continuous language from non-invasive recordings
would have many scientific and practical applications. Currently, however, non-invasive …

The science of detecting LLM-generated text

R Tang, YN Chuang, X Hu - Communications of the ACM, 2024 - dl.acm.org
ACM: Digital Library: Communications of the ACM ACM Digital Library Communications of the
ACM Volume 67, Number 4 (2024), Pages 50-59 The Science of Detecting LLM-Generated Text …

Calibrating sequence likelihood improves conditional language generation

Y Zhao, M Khalman, R Joshi, S Narayan… - arxiv preprint arxiv …, 2022 - arxiv.org
Conditional language models are predominantly trained with maximum likelihood estimation
(MLE), giving probability mass to sparsely observed target sequences. While MLE trained …

Amortizing intractable inference in large language models

EJ Hu, M Jain, E Elmoznino, Y Kaddar, G Lajoie… - arxiv preprint arxiv …, 2023 - arxiv.org
Autoregressive large language models (LLMs) compress knowledge from their training data
through next-token conditional distributions. This limits tractable querying of this knowledge …

Deep learning driven biosynthetic pathways navigation for natural products with BioNavi-NP

S Zheng, T Zeng, C Li, B Chen, CW Coley… - Nature …, 2022 - nature.com
The complete biosynthetic pathways are unknown for most natural products (NPs), it is thus
valuable to make computer-aided bio-retrosynthesis predictions. Here, a navigable and user …

Cycle: Learning to self-refine the code generation

Y Ding, MJ Min, G Kaiser, B Ray - Proceedings of the ACM on …, 2024 - dl.acm.org
Pre-trained code language models have achieved promising performance in code
generation and improved the programming efficiency of human developers. However, their …

Alphamath almost zero: process supervision without process

G Chen, M Liao, C Li, K Fan - arxiv preprint arxiv:2405.03553, 2024 - arxiv.org
Although recent advancements in large language models (LLMs) have significantly
improved their performance on various tasks, they still face challenges with complex and …

[HTML][HTML] Decoding methods in neural language generation: a survey

S Zarrieß, H Voigt, S Schüz - Information, 2021 - mdpi.com
Neural encoder-decoder models for language generation can be trained to predict words
directly from linguistic or non-linguistic inputs. When generating with these so-called end-to …

The transformer network for the traveling salesman problem

X Bresson, T Laurent - arxiv preprint arxiv:2103.03012, 2021 - arxiv.org
The Traveling Salesman Problem (TSP) is the most popular and most studied combinatorial
problem, starting with von Neumann in 1951. It has driven the discovery of several …