[HTML][HTML] A survey of GPT-3 family large language models including ChatGPT and GPT-4

KS Kalyan - Natural Language Processing Journal, 2024 - Elsevier
Large language models (LLMs) are a special class of pretrained language models (PLMs)
obtained by scaling model size, pretraining corpus and computation. LLMs, because of their …

C-pack: Packed resources for general chinese embeddings

S **ao, Z Liu, P Zhang, N Muennighoff, D Lian… - Proceedings of the 47th …, 2024 - dl.acm.org
We introduce C-Pack, a package of resources that significantly advances the field of general
text embeddings for Chinese. C-Pack includes three critical resources. 1) C-MTP is a …

Large language models for information retrieval: A survey

Y Zhu, H Yuan, S Wang, J Liu, W Liu, C Deng… - arxiv preprint arxiv …, 2023 - arxiv.org
As a primary means of information acquisition, information retrieval (IR) systems, such as
search engines, have integrated themselves into our daily lives. These systems also serve …

Evaluating large language models at evaluating instruction following

Z Zeng, J Yu, T Gao, Y Meng, T Goyal… - arxiv preprint arxiv …, 2023 - arxiv.org
As research in large language models (LLMs) continues to accelerate, LLM-based
evaluation has emerged as a scalable and cost-effective alternative to human evaluations …

Improving text embeddings with large language models

L Wang, N Yang, X Huang, L Yang… - arxiv preprint arxiv …, 2023 - arxiv.org
In this paper, we introduce a novel and simple method for obtaining high-quality text
embeddings using only synthetic data and less than 1k training steps. Unlike existing …

Replug: Retrieval-augmented black-box language models

W Shi, S Min, M Yasunaga, M Seo, R James… - arxiv preprint arxiv …, 2023 - arxiv.org
We introduce REPLUG, a retrieval-augmented language modeling framework that treats the
language model (LM) as a black box and augments it with a tuneable retrieval model. Unlike …

Exploring the benefits of training expert language models over instruction tuning

J Jang, S Kim, S Ye, D Kim… - International …, 2023 - proceedings.mlr.press
Abstract Recently, Language Models (LMs) instruction-tuned on multiple tasks, also known
as multitask-prompted fine-tuning (MT), have shown capabilities to generalize to unseen …

Representation learning with large language models for recommendation

X Ren, W Wei, L **a, L Su, S Cheng, J Wang… - Proceedings of the …, 2024 - dl.acm.org
Recommender systems have seen significant advancements with the influence of deep
learning and graph neural networks, particularly in capturing complex user-item …

Augmenting interpretable models with large language models during training

C Singh, A Askari, R Caruana, J Gao - Nature Communications, 2023 - nature.com
Recent large language models (LLMs), such as ChatGPT, have demonstrated remarkable
prediction performance for a growing array of tasks. However, their proliferation into high …

Uniir: Training and benchmarking universal multimodal information retrievers

C Wei, Y Chen, H Chen, H Hu, G Zhang, J Fu… - … on Computer Vision, 2024 - Springer
Existing information retrieval (IR) models often assume a homogeneous format, limiting their
applicability to diverse user needs, such as searching for images with text descriptions …