Phi-3 technical report: A highly capable language model locally on your phone

M Abdin, J Aneja, H Awadalla, A Awadallah… - arxiv preprint arxiv …, 2024 - arxiv.org
We introduce phi-3-mini, a 3.8 billion parameter language model trained on 3.3 trillion
tokens, whose overall performance, as measured by both academic benchmarks and …

The life cycle of large language models in education: A framework for understanding sources of bias

J Lee, Y Hicke, R Yu, C Brooks… - British Journal of …, 2024 - Wiley Online Library
Large language models (LLMs) are increasingly adopted in educational contexts to provide
personalized support to students and teachers. The unprecedented capacity of LLM‐based …

Siren's song in the AI ocean: a survey on hallucination in large language models

Y Zhang, Y Li, L Cui, D Cai, L Liu, T Fu… - arxiv preprint arxiv …, 2023 - arxiv.org
While large language models (LLMs) have demonstrated remarkable capabilities across a
range of downstream tasks, a significant concern revolves around their propensity to exhibit …

Towards trustworthy LLMs: a review on debiasing and dehallucinating in large language models

Z Lin, S Guan, W Zhang, H Zhang, Y Li… - Artificial Intelligence …, 2024 - Springer
Recently, large language models (LLMs) have attracted considerable attention due to their
remarkable capabilities. However, LLMs' generation of biased or hallucinatory content …

Llamafactory: Unified efficient fine-tuning of 100+ language models

Y Zheng, R Zhang, J Zhang, Y Ye, Z Luo… - arxiv preprint arxiv …, 2024 - arxiv.org
Efficient fine-tuning is vital for adapting large language models (LLMs) to downstream tasks.
However, it requires non-trivial efforts to implement these methods on different models. We …

Multimodal foundation models: From specialists to general-purpose assistants

C Li, Z Gan, Z Yang, J Yang, L Li… - … and Trends® in …, 2024 - nowpublishers.com
Neural compression is the application of neural networks and other machine learning
methods to data compression. Recent advances in statistical machine learning have opened …

Llava-phi: Efficient multi-modal assistant with small language model

Y Zhu, M Zhu, N Liu, Z Xu, Y Peng - … of the 1st International Workshop on …, 2024 - dl.acm.org
In this paper, we introduce LLaVA-φ (LLaVA-Phi), an efficient multi-modal assistant that
harnesses the power of the recently advanced small language model, Phi-2, to facilitate …

Large language models: A survey

S Minaee, T Mikolov, N Nikzad, M Chenaghlu… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) have drawn a lot of attention due to their strong
performance on a wide range of natural language tasks, since the release of ChatGPT in …

Show-o: One single transformer to unify multimodal understanding and generation

J **e, W Mao, Z Bai, DJ Zhang, W Wang, KQ Lin… - arxiv preprint arxiv …, 2024 - arxiv.org
We present a unified transformer, ie, Show-o, that unifies multimodal understanding and
generation. Unlike fully autoregressive models, Show-o unifies autoregressive and …

On llms-driven synthetic data generation, curation, and evaluation: A survey

L Long, R Wang, R **ao, J Zhao, X Ding… - arxiv preprint arxiv …, 2024 - arxiv.org
Within the evolving landscape of deep learning, the dilemma of data quantity and quality has
been a long-standing problem. The recent advent of Large Language Models (LLMs) offers …