Recent advances in recurrent neural networks
Recurrent neural networks (RNNs) are capable of learning features and long term
dependencies from sequential and time-series data. The RNNs have a stack of non-linear …
dependencies from sequential and time-series data. The RNNs have a stack of non-linear …
A primer on neural network models for natural language processing
Y Goldberg - Journal of Artificial Intelligence Research, 2016 - jair.org
Over the past few years, neural networks have re-emerged as powerful machine-learning
models, yielding state-of-the-art results in fields such as image recognition and speech …
models, yielding state-of-the-art results in fields such as image recognition and speech …
[LIBRO][B] Neural network methods in natural language processing
Y Goldberg - 2017 - books.google.com
Neural networks are a family of powerful machine learning models and this book focuses on
their application to natural language data. The first half of the book (Parts I and II) covers the …
their application to natural language data. The first half of the book (Parts I and II) covers the …
Enriching word vectors with subword information
Continuous word representations, trained on large unlabeled corpora are useful for many
natural language processing tasks. Popular models that learn such representations ignore …
natural language processing tasks. Popular models that learn such representations ignore …
Character-aware neural language models
We describe a simple neural language model that relies only on character-level inputs.
Predictions are still made at the word-level. Our model employs a convolutional neural …
Predictions are still made at the word-level. Our model employs a convolutional neural …
75 languages, 1 model: Parsing universal dependencies universally
We present UDify, a multilingual multi-task model capable of accurately predicting universal
part-of-speech, morphological features, lemmas, and dependency trees simultaneously for …
part-of-speech, morphological features, lemmas, and dependency trees simultaneously for …
Globally normalized transition-based neural networks
We introduce a globally normalized transition-based neural network model that achieves
state-of-the-art part-of-speech tagging, dependency parsing and sentence compression …
state-of-the-art part-of-speech tagging, dependency parsing and sentence compression …
Finding function in form: Compositional character models for open vocabulary word representation
We introduce a model for constructing vector representations of words by composing
characters using bidirectional LSTMs. Relative to traditional word representation models that …
characters using bidirectional LSTMs. Relative to traditional word representation models that …
Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss
Bidirectional long short-term memory (bi-LSTM) networks have recently proven successful
for various NLP sequence modeling tasks, but little is known about their reliance to input …
for various NLP sequence modeling tasks, but little is known about their reliance to input …
Achieving open vocabulary neural machine translation with hybrid word-character models
Nearly all previous work on neural machine translation (NMT) has used quite restricted
vocabularies, perhaps with a subsequent method to patch in unknown words. This paper …
vocabularies, perhaps with a subsequent method to patch in unknown words. This paper …