Neural transfer learning for repairing security vulnerabilities in c code

Z Chen, S Kommrusch… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
In this paper, we address the problem of automatic repair of software vulnerabilities with
deep learning. The major problem with data-driven vulnerability repair is that the few …

Linguistics-based formalization of the antibody language as a basis for antibody language models

MH Vu, PA Robert, R Akbar, B Swiatczak… - Nature Computational …, 2024 - nature.com
Apparent parallels between natural language and antibody sequences have led to a surge
in deep language models applied to antibody sequences for predicting cognate antigen …

Good-enough compositional data augmentation

J Andreas - arxiv preprint arxiv:1904.09545, 2019 - arxiv.org
We propose a simple data augmentation protocol aimed at providing a compositional
inductive bias in conditional and unconditional sequence models. Under this protocol …

Participatory research for low-resourced machine translation: A case study in african languages

W Nekoto, V Marivate, T Matsila, T Fasubaa… - arxiv preprint arxiv …, 2020 - arxiv.org
Research in NLP lacks geographic diversity, and the question of how NLP can be scaled to
low-resourced languages has not yet been adequately solved." Low-resourced"-ness is a …

Cross-lingual alignment of contextual word embeddings, with applications to zero-shot dependency parsing

T Schuster, O Ram, R Barzilay, A Globerson - arxiv preprint arxiv …, 2019 - arxiv.org
We introduce a novel method for multilingual transfer that utilizes deep contextual
embeddings, pretrained in an unsupervised fashion. While contextual embeddings have …

Dual adversarial neural transfer for low-resource named entity recognition

JT Zhou, H Zhang, D **, H Zhu, M Fang… - Proceedings of the …, 2019 - aclanthology.org
We propose a new neural transfer method termed Dual Adversarial Transfer Network
(DATNet) for addressing low-resource Named Entity Recognition (NER). Specifically, two …

A survey on transfer learning in natural language processing

Z Alyafeai, MS AlShaibani, I Ahmad - arxiv preprint arxiv:2007.04239, 2020 - arxiv.org
Deep learning models usually require a huge amount of data. However, these large
datasets are not always attainable. This is common in many challenging NLP tasks …