Challenges and applications of large language models

J Kaddour, J Harris, M Mozes, H Bradley… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) went from non-existent to ubiquitous in the machine
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …

When Can LLMs Actually Correct Their Own Mistakes? A Critical Survey of Self-Correction of LLMs

R Kamoi, Y Zhang, N Zhang, J Han… - Transactions of the …, 2024 - direct.mit.edu
Self-correction is an approach to improving responses from large language models (LLMs)
by refining the responses using LLMs during inference. Prior work has proposed various self …

AI generates covertly racist decisions about people based on their dialect

V Hofmann, PR Kalluri, D Jurafsky, S King - Nature, 2024 - nature.com
Hundreds of millions of people now interact with language models, with uses ranging from
help with writing, to informing hiring decisions. However, these language models are known …

Cognitive architectures for language agents

TR Sumers, S Yao, K Narasimhan… - arxiv preprint arxiv …, 2023 - arxiv.org
Recent efforts have incorporated large language models (LLMs) with external resources (eg,
the Internet) or internal control flows (eg, prompt chaining) for tasks requiring grounding or …

Large language models to identify social determinants of health in electronic health records

M Guevara, S Chen, S Thomas, TL Chaunzwa… - NPJ digital …, 2024 - nature.com
Social determinants of health (SDoH) play a critical role in patient outcomes, yet their
documentation is often missing or incomplete in the structured data of electronic health …

Openfedllm: Training large language models on decentralized private data via federated learning

R Ye, W Wang, J Chai, D Li, Z Li, Y Xu, Y Du… - Proceedings of the 30th …, 2024 - dl.acm.org
Trained on massive publicly available data, large language models (LLMs) have
demonstrated tremendous success across various fields. While more data contributes to …

Do large language models have a legal duty to tell the truth?

S Wachter, B Mittelstadt… - Royal Society Open …, 2024 - royalsocietypublishing.org
Careless speech is a new type of harm created by large language models (LLM) that poses
cumulative, long-term risks to science, education and shared social truth in democratic …

AI can help humans find common ground in democratic deliberation

MH Tessler, MA Bakker, D Jarrett, H Sheahan… - Science, 2024 - science.org
Finding agreement through a free exchange of views is often difficult. Collective deliberation
can be slow, difficult to scale, and unequally attentive to different voices. In this study, we …

Fairness in large language models: A taxonomic survey

Z Chu, Z Wang, W Zhang - ACM SIGKDD explorations newsletter, 2024 - dl.acm.org
Large Language Models (LLMs) have demonstrated remarkable success across various
domains. However, despite their promising performance in numerous real-world …

Personalized soups: Personalized large language model alignment via post-hoc parameter merging

J Jang, S Kim, BY Lin, Y Wang, J Hessel… - arxiv preprint arxiv …, 2023 - arxiv.org
While Reinforcement Learning from Human Feedback (RLHF) aligns Large Language
Models (LLMs) with general, aggregate human preferences, it is suboptimal for learning …