Knowledge graphs meet multi-modal learning: A comprehensive survey

Z Chen, Y Zhang, Y Fang, Y Geng, L Guo… - arxiv preprint arxiv …, 2024 - arxiv.org
Knowledge Graphs (KGs) play a pivotal role in advancing various AI applications, with the
semantic web community's exploration into multi-modal dimensions unlocking new avenues …

A survey on mixture of experts

W Cai, J Jiang, F Wang, J Tang, S Kim… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) have garnered unprecedented advancements across
diverse fields, ranging from natural language processing to computer vision and beyond …

Magis: Llm-based multi-agent framework for github issue resolution

W Tao, Y Zhou, Y Wang, W Zhang, H Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
In software evolution, resolving the emergent issues within GitHub repositories is a complex
challenge that involves not only the incorporation of new code but also the maintenance of …

Conflictbank: A benchmark for evaluating the influence of knowledge conflicts in llm

Z Su, J Zhang, X Qu, T Zhu, Y Li, J Sun, J Li… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) have achieved impressive advancements across numerous
disciplines, yet the critical issue of knowledge conflicts, a major source of hallucinations, has …

A Survey: Collaborative Hardware and Software Design in the Era of Large Language Models

C Guo, F Cheng, Z Du, J Kiessling, J Ku, S Li… - arxiv preprint arxiv …, 2024 - arxiv.org
The rapid development of large language models (LLMs) has significantly transformed the
field of artificial intelligence, demonstrating remarkable capabilities in natural language …

Fusion: Helpful, Harmless, Honest Fusion of Aligned LLMs

SF Tekin, F Ilhan, T Huang, S Hu, Z Yahn… - arxiv preprint arxiv …, 2024 - arxiv.org
Alignment of pretrained LLMs using instruction-based datasets is critical for creating fine-
tuned models that reflect human preference. A growing number of alignment-based fine …

On Giant's Shoulders: Effortless Weak to Strong by Dynamic Logits Fusion

C Fan, Z Lu, W Wei, J Tian, X Qu, D Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
Efficient fine-tuning of large language models for task-specific applications is imperative, yet
the vast number of parameters in these models makes their training increasingly …

Moe++: Accelerating mixture-of-experts methods with zero-computation experts

P **, B Zhu, L Yuan, S Yan - arxiv preprint arxiv:2410.07348, 2024 - arxiv.org
In this work, we aim to simultaneously enhance the effectiveness and efficiency of Mixture-of-
Experts (MoE) methods. To achieve this, we propose MoE++, a general and heterogeneous …

DLO: Dynamic Layer Operation for Efficient Vertical Scaling of LLMs

Z Tan, D Dong, X Zhao, J Peng, Y Cheng… - arxiv preprint arxiv …, 2024 - arxiv.org
In this paper, we introduce Dynamic Layer Operations (DLO), a novel approach for vertically
scaling transformer-based Large Language Models (LLMs) by dynamically expanding …

Performance Law of Large Language Models

C Wu, R Tang - arxiv preprint arxiv:2408.09895, 2024 - arxiv.org
Guided by the belief of the scaling law, large language models (LLMs) have achieved
impressive performance in recent years. However, scaling law only gives a qualitative …