Retrieval-augmented generation for large language models: A survey
Y Gao, Y **s. We investigate two setups-ICL with flipped labels and ICL with …
Focused transformer: Contrastive training for context scaling
Large language models have an exceptional capability to incorporate new information in a
contextual manner. However, the full potential of such an approach is often restrained due to …
contextual manner. However, the full potential of such an approach is often restrained due to …
Finetuned language models are zero-shot learners
Multitask prompted training enables zero-shot task generalization
Large language models have recently been shown to attain reasonable zero-shot
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that …
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that …
[PDF][PDF] Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
N Reimers - arxiv preprint arxiv:1908.10084, 2019 - fq.pkwyx.com
BERT (Devlin et al., 2018) and RoBERTa (Liu et al., 2019) has set a new state-of-the-art
performance on sentence-pair regression tasks like semantic textual similarity (STS) …
performance on sentence-pair regression tasks like semantic textual similarity (STS) …
Bae: Bert-based adversarial examples for text classification
Modern text classification models are susceptible to adversarial examples, perturbed
versions of the original text indiscernible by humans which get misclassified by the model …
versions of the original text indiscernible by humans which get misclassified by the model …