Federated learning from pre-trained models: A contrastive learning approach
Federated Learning (FL) is a machine learning paradigm that allows decentralized clients to
learn collaboratively without sharing their private data. However, excessive computation and …
learn collaboratively without sharing their private data. However, excessive computation and …
Sentence-t5: Scalable sentence encoders from pre-trained text-to-text models
We provide the first exploration of sentence embeddings from text-to-text transformers (T5).
Sentence embeddings are broadly useful for language processing tasks. While T5 achieves …
Sentence embeddings are broadly useful for language processing tasks. While T5 achieves …
Fedproto: Federated prototype learning across heterogeneous clients
Heterogeneity across clients in federated learning (FL) usually hinders the optimization
convergence and generalization performance when the aggregation of clients' knowledge …
convergence and generalization performance when the aggregation of clients' knowledge …
[LIBRO][B] Pretrained transformers for text ranking: Bert and beyond
The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in
response to a query. Although the most common formulation of text ranking is search …
response to a query. Although the most common formulation of text ranking is search …
A brief overview of universal sentence representation methods: A linguistic view
How to transfer the semantic information in a sentence to a computable numerical
embedding form is a fundamental problem in natural language processing. An informative …
embedding form is a fundamental problem in natural language processing. An informative …
BLEURT: Learning robust metrics for text generation
Text generation has made significant advances in the last few years. Yet, evaluation metrics
have lagged behind, as the most popular choices (eg, BLEU and ROUGE) may correlate …
have lagged behind, as the most popular choices (eg, BLEU and ROUGE) may correlate …
Parameter-efficient transfer learning with diff pruning
While task-specific finetuning of pretrained networks has led to significant empirical
advances in NLP, the large size of networks makes finetuning difficult to deploy in multi-task …
advances in NLP, the large size of networks makes finetuning difficult to deploy in multi-task …
Aligning ai with shared human values
We show how to assess a language model's knowledge of basic concepts of morality. We
introduce the ETHICS dataset, a new benchmark that spans concepts in justice, well-being …
introduce the ETHICS dataset, a new benchmark that spans concepts in justice, well-being …
Pretrained transformers improve out-of-distribution robustness
Although pretrained Transformers such as BERT achieve high accuracy on in-distribution
examples, do they generalize to new distributions? We systematically measure out-of …
examples, do they generalize to new distributions? We systematically measure out-of …
Transfer learning in natural language processing
The classic supervised machine learning paradigm is based on learning in isolation, a
single predictive model for a task using a single dataset. This approach requires a large …
single predictive model for a task using a single dataset. This approach requires a large …