Retrieval augmented generation (rag) and beyond: A comprehensive survey on how to make your llms use external data more wisely

S Zhao, Y Yang, Z Wang, Z He, LK Qiu… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) augmented with external data have demonstrated
remarkable capabilities in completing real-world tasks. Techniques for integrating external …

Multilingual machine translation with large language models: Empirical results and analysis

W Zhu, H Liu, Q Dong, J Xu, S Huang, L Kong… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) have demonstrated remarkable potential in handling
multilingual machine translation (MMT). In this paper, we systematically investigate the …

Os-copilot: Towards generalist computer agents with self-improvement

Z Wu, C Han, Z Ding, Z Weng, Z Liu, S Yao… - arxiv preprint arxiv …, 2024 - arxiv.org
Autonomous interaction with the computer has been a longstanding challenge with great
potential, and the recent proliferation of large language models (LLMs) has markedly …

Small models are valuable plug-ins for large language models

C Xu, Y Xu, S Wang, Y Liu, C Zhu… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) such as GPT-3 and GPT-4 are powerful but their weights are
often publicly unavailable and their immense sizes make the models difficult to be tuned with …

Corex: Pushing the boundaries of complex reasoning through multi-model collaboration

Q Sun, Z Yin, X Li, Z Wu, X Qiu, L Kong - arxiv preprint arxiv:2310.00280, 2023 - arxiv.org
Large Language Models (LLMs) are evolving at an unprecedented pace and have exhibited
considerable capability in the realm of natural language processing (NLP) with world …

Revisiting demonstration selection strategies in in-context learning

K Peng, L Ding, Y Yuan, X Liu, M Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) have shown an impressive ability to perform a wide range of
tasks using in-context learning (ICL), where a few examples are used to describe a task to …

MetaAdapt: Domain adaptive few-shot misinformation detection via meta learning

Z Yue, H Zeng, Y Zhang, L Shang, D Wang - arxiv preprint arxiv …, 2023 - arxiv.org
With emerging topics (eg, COVID-19) on social media as a source for the spreading
misinformation, overcoming the distributional shifts between the original training domain (ie …

Debiasing multimodal sarcasm detection with contrastive learning

M Jia, C **e, L **g - Proceedings of the AAAI Conference on Artificial …, 2024 - ojs.aaai.org
Despite commendable achievements made by existing work, prevailing multimodal sarcasm
detection studies rely more on textual content over visual information. It unavoidably induces …

Forward-backward reasoning in large language models for mathematical verification

W Jiang, H Shi, L Yu, Z Liu, Y Zhang, Z Li… - arxiv preprint arxiv …, 2023 - arxiv.org
Self-Consistency samples diverse reasoning chains with answers and chooses the final
answer by majority voting. It is based on forward reasoning and cannot further improve …

Llms for low resource languages in multilingual, multimodal and dialectal settings

F Alam, SA Chowdhury, S Boughorbel… - Proceedings of the …, 2024 - aclanthology.org
The recent breakthroughs in Artificial Intelligence (AI) can be attributed to the remarkable
performance of Large Language Models (LLMs) across a spectrum of research areas (eg …