A tutorial on deep latent variable models of natural language

Y Kim, S Wiseman, AM Rush - arxiv preprint arxiv:1812.06834, 2018 - arxiv.org
There has been much recent, exciting work on combining the complementary strengths of
latent variable models and deep learning. Latent variable modeling makes it easy to …

Sequence-to-sequence learning with latent neural grammars

Y Kim - Advances in Neural Information Processing …, 2021 - proceedings.neurips.cc
Sequence-to-sequence learning with neural networks has become the de facto standard for
sequence modeling. This approach typically models the local distribution over the next …

Visually grounded neural syntax acquisition

H Shi, J Mao, K Gimpel, K Livescu - arxiv preprint arxiv:1906.02890, 2019 - arxiv.org
We present the Visually Grounded Neural Syntax Learner (VG-NSL), an approach for
learning syntactic representations and structures without any explicit supervision. The model …

StructFormer: Joint unsupervised induction of dependency and constituency structure from masked language modeling

Y Shen, Y Tay, C Zheng, D Bahri, D Metzler… - arxiv preprint arxiv …, 2020 - arxiv.org
There are two major classes of natural language grammar--the dependency grammar that
models one-to-one correspondences between words and the constituency grammar that …

Semantic role labeling as dependency parsing: Exploring latent tree structures inside arguments

Y Zhang, Q **a, S Zhou, Y Jiang, G Fu… - arxiv preprint arxiv …, 2021 - arxiv.org
Semantic role labeling (SRL) is a fundamental yet challenging task in the NLP community.
Recent works of SRL mainly fall into two lines: 1) BIO-based; 2) span-based. Despite …

[PDF][PDF] Unsupervised vision-language grammar induction with shared structure modeling

B Wan, W Han, Z Zheng, T Tuytelaars - Proceedings ICLR 2022, 2022 - lirias.kuleuven.be
We introduce a new task, unsupervised vision-language (VL) grammar induction. Given an
image-caption pair, the goal is to extract a shared hierarchical structure for both image and …

Unsupervised learning of syntactic structure with invertible neural projections

J He, G Neubig, T Berg-Kirkpatrick - arxiv preprint arxiv:1808.09111, 2018 - arxiv.org
Unsupervised learning of syntactic structure is typically performed using generative models
with discrete latent variables and multinomial parameters. In most cases, these models have …

Hierarchical phrase-based sequence-to-sequence learning

B Wang, I Titov, J Andreas, Y Kim - arxiv preprint arxiv:2211.07906, 2022 - arxiv.org
We describe a neural transducer that maintains the flexibility of standard sequence-to-
sequence (seq2seq) models while incorporating hierarchical phrases as a source of …

Neural bi-lexicalized PCFG induction

S Yang, Y Zhao, K Tu - arxiv preprint arxiv:2105.15021, 2021 - arxiv.org
Neural lexicalized PCFGs (L-PCFGs) have been shown effective in grammar induction.
However, to reduce computational complexity, they make a strong independence …

Scaling hidden Markov language models

JT Chiu, AM Rush - arxiv preprint arxiv:2011.04640, 2020 - arxiv.org
The hidden Markov model (HMM) is a fundamental tool for sequence modeling that cleanly
separates the hidden state from the emission structure. However, this separation makes it …