Federated learning from pre-trained models: A contrastive learning approach

Y Tan, G Long, J Ma, L Liu, T Zhou… - Advances in neural …, 2022 - proceedings.neurips.cc
Federated Learning (FL) is a machine learning paradigm that allows decentralized clients to
learn collaboratively without sharing their private data. However, excessive computation and …

Sentence-t5: Scalable sentence encoders from pre-trained text-to-text models

J Ni, GH Abrego, N Constant, J Ma, KB Hall… - arxiv preprint arxiv …, 2021 - arxiv.org
We provide the first exploration of sentence embeddings from text-to-text transformers (T5).
Sentence embeddings are broadly useful for language processing tasks. While T5 achieves …

Fedproto: Federated prototype learning across heterogeneous clients

Y Tan, G Long, L Liu, T Zhou, Q Lu, J Jiang… - Proceedings of the …, 2022 - ojs.aaai.org
Heterogeneity across clients in federated learning (FL) usually hinders the optimization
convergence and generalization performance when the aggregation of clients' knowledge …

[LIBRO][B] Pretrained transformers for text ranking: Bert and beyond

J Lin, R Nogueira, A Yates - 2022 - books.google.com
The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in
response to a query. Although the most common formulation of text ranking is search …

A brief overview of universal sentence representation methods: A linguistic view

R Li, X Zhao, MF Moens - ACM Computing Surveys (CSUR), 2022 - dl.acm.org
How to transfer the semantic information in a sentence to a computable numerical
embedding form is a fundamental problem in natural language processing. An informative …

BLEURT: Learning robust metrics for text generation

T Sellam, D Das, AP Parikh - arxiv preprint arxiv:2004.04696, 2020 - arxiv.org
Text generation has made significant advances in the last few years. Yet, evaluation metrics
have lagged behind, as the most popular choices (eg, BLEU and ROUGE) may correlate …

Parameter-efficient transfer learning with diff pruning

D Guo, AM Rush, Y Kim - arxiv preprint arxiv:2012.07463, 2020 - arxiv.org
While task-specific finetuning of pretrained networks has led to significant empirical
advances in NLP, the large size of networks makes finetuning difficult to deploy in multi-task …

Aligning ai with shared human values

D Hendrycks, C Burns, S Basart, A Critch, J Li… - arxiv preprint arxiv …, 2020 - arxiv.org
We show how to assess a language model's knowledge of basic concepts of morality. We
introduce the ETHICS dataset, a new benchmark that spans concepts in justice, well-being …

Pretrained transformers improve out-of-distribution robustness

D Hendrycks, X Liu, E Wallace, A Dziedzic… - arxiv preprint arxiv …, 2020 - arxiv.org
Although pretrained Transformers such as BERT achieve high accuracy on in-distribution
examples, do they generalize to new distributions? We systematically measure out-of …

Transfer learning in natural language processing

S Ruder, ME Peters, S Swayamdipta… - Proceedings of the 2019 …, 2019 - aclanthology.org
The classic supervised machine learning paradigm is based on learning in isolation, a
single predictive model for a task using a single dataset. This approach requires a large …