Evaluating large language models: A comprehensive survey

Z Guo, R **, C Liu, Y Huang, D Shi, L Yu, Y Liu… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) have demonstrated remarkable capabilities across a broad
spectrum of tasks. They have attracted significant attention and been deployed in numerous …

The evolution of distributed systems for graph neural networks and their origin in graph processing and deep learning: A survey

J Vatter, R Mayer, HA Jacobsen - ACM Computing Surveys, 2023 - dl.acm.org
Graph neural networks (GNNs) are an emerging research field. This specialized deep
neural network architecture is capable of processing graph structured data and bridges the …

Gpt-4 technical report

J Achiam, S Adler, S Agarwal, L Ahmad… - arxiv preprint arxiv …, 2023 - arxiv.org
We report the development of GPT-4, a large-scale, multimodal model which can accept
image and text inputs and produce text outputs. While less capable than humans in many …

Openassistant conversations-democratizing large language model alignment

A Köpf, Y Kilcher, D Von Rütte… - Advances in …, 2023 - proceedings.neurips.cc
Aligning large language models (LLMs) with human preferences has proven to drastically
improve usability and has driven rapid adoption as demonstrated by ChatGPT. Alignment …

Are aligned neural networks adversarially aligned?

N Carlini, M Nasr… - Advances in …, 2023 - proceedings.neurips.cc
Large language models are now tuned to align with the goals of their creators, namely to be"
helpful and harmless." These models should respond helpfully to user questions, but refuse …

Should chatgpt be biased? challenges and risks of bias in large language models

E Ferrara - arxiv preprint arxiv:2304.03738, 2023 - arxiv.org
As the capabilities of generative language models continue to advance, the implications of
biases ingrained within these models have garnered increasing attention from researchers …

[PDF][PDF] Trustllm: Trustworthiness in large language models

L Sun, Y Huang, H Wang, S Wu, Q Zhang… - arxiv preprint arxiv …, 2024 - mosis.eecs.utk.edu
Large language models (LLMs), exemplified by ChatGPT, have gained considerable
attention for their excellent natural language processing capabilities. Nonetheless, these …

Red teaming language models to reduce harms: Methods, scaling behaviors, and lessons learned

D Ganguli, L Lovitt, J Kernion, A Askell, Y Bai… - arxiv preprint arxiv …, 2022 - arxiv.org
We describe our early efforts to red team language models in order to simultaneously
discover, measure, and attempt to reduce their potentially harmful outputs. We make three …

From pretraining data to language models to downstream tasks: Tracking the trails of political biases leading to unfair NLP models

S Feng, CY Park, Y Liu, Y Tsvetkov - arxiv preprint arxiv:2305.08283, 2023 - arxiv.org
Language models (LMs) are pretrained on diverse data sources, including news, discussion
forums, books, and online encyclopedias. A significant portion of this data includes opinions …

Unified concept editing in diffusion models

R Gandikota, H Orgad, Y Belinkov… - Proceedings of the …, 2024 - openaccess.thecvf.com
Text-to-image models suffer from various safety issues that may limit their suitability for
deployment. Previous methods have separately addressed individual issues of bias …