Codexgraph: Bridging large language models and code repositories via code graph databases

X Liu, B Lan, Z Hu, Y Liu, Z Zhang, F Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) excel in stand-alone code tasks like HumanEval and
MBPP, but struggle with handling entire code repositories. This challenge has prompted …

SysBench: Can Large Language Models Follow System Messages?

Y Qin, T Zhang, Y Shen, W Luo, H Sun, Y Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) have become instrumental across various applications,
with the customization of these models to specific scenarios becoming increasingly critical …

Critic-cot: Boosting the reasoning abilities of large language model via chain-of-thoughts critic

X Zheng, J Lou, B Cao, X Wen, Y Ji, H Lin, Y Lu… - arxiv preprint arxiv …, 2024 - arxiv.org
Self-critic has become a crucial mechanism for enhancing the reasoning performance of
LLMs. However, current approaches mainly involve basic prompts for intuitive instance-level …

Scilitllm: How to adapt llms for scientific literature understanding

S Li, J Huang, J Zhuang, Y Shi, X Cai, M Xu… - arxiv preprint arxiv …, 2024 - arxiv.org
Scientific literature understanding is crucial for extracting targeted information and garnering
insights, thereby significantly advancing scientific discovery. Despite the remarkable …

Training language models to critique with multi-agent feedback

T Lan, W Zhang, C Lyu, S Li, C Xu, H Huang… - arxiv preprint arxiv …, 2024 - arxiv.org
Critique ability, a meta-cognitive capability of humans, presents significant challenges for
LLMs to improve. Recent works primarily rely on supervised fine-tuning (SFT) using critiques …

LLMCO2: Advancing Accurate Carbon Footprint Prediction for LLM Inferences

Z Fu, F Chen, S Zhou, H Li, L Jiang - arxiv preprint arxiv:2410.02950, 2024 - arxiv.org
Throughout its lifecycle, a large language model (LLM) generates a substantially larger
carbon footprint during inference than training. LLM inference requests vary in batch size …

A Comprehensive Survey of Direct Preference Optimization: Datasets, Theories, Variants, and Applications

W **ao, Z Wang, L Gan, S Zhao, W He, LA Tuan… - arxiv preprint arxiv …, 2024 - arxiv.org
With the rapid advancement of large language models (LLMs), aligning policy models with
human preferences has become increasingly critical. Direct Preference Optimization (DPO) …

PoisonBench: Assessing Large Language Model Vulnerability to Data Poisoning

T Fu, M Sharma, P Torr, SB Cohen, D Krueger… - arxiv preprint arxiv …, 2024 - arxiv.org
Preference learning is a central component for aligning current LLMs, but this process can
be vulnerable to data poisoning attacks. To address this concern, we introduce …

TensorOpera Router: A Multi-Model Router for Efficient LLM Inference

D Stripelis, Z Hu, J Zhang, Z Xu, AD Shah, H **… - arxiv preprint arxiv …, 2024 - arxiv.org
With the rapid growth of Large Language Models (LLMs) across various domains, numerous
new LLMs have emerged, each possessing domain-specific expertise. This proliferation has …

Aligning large language models via self-steering optimization

H **ang, B Yu, H Lin, K Lu, Y Lu, X Han, L Sun… - arxiv preprint arxiv …, 2024 - arxiv.org
Automated alignment develops alignment systems with minimal human intervention. The
key to automated alignment lies in providing learnable and accurate preference signals for …