Chatgpt needs spade (sustainability, privacy, digital divide, and ethics) evaluation: A review

SA Khowaja, P Khuwaja, K Dev, W Wang… - Cognitive …, 2024 - Springer
ChatGPT is another large language model (LLM) vastly available for the consumers on their
devices but due to its performance and ability to converse effectively, it has gained a huge …

[HTML][HTML] When llms meet cybersecurity: A systematic literature review

J Zhang, H Bu, H Wen, Y Liu, H Fei… - …, 2025 - cybersecurity.springeropen.com
The rapid development of large language models (LLMs) has opened new avenues across
various fields, including cybersecurity, which faces an evolving threat landscape and …

The llama 3 herd of models

A Dubey, A Jauhri, A Pandey, A Kadian… - arxiv preprint arxiv …, 2024 - arxiv.org
Modern artificial intelligence (AI) systems are powered by foundation models. This paper
presents a new set of foundation models, called Llama 3. It is a herd of language models …

Harmbench: A standardized evaluation framework for automated red teaming and robust refusal

M Mazeika, L Phan, X Yin, A Zou, Z Wang, N Mu… - arxiv preprint arxiv …, 2024 - arxiv.org
Automated red teaming holds substantial promise for uncovering and mitigating the risks
associated with the malicious use of large language models (LLMs), yet the field lacks a …

On prompt-driven safeguarding for large language models

C Zheng, F Yin, H Zhou, F Meng, J Zhou… - arxiv preprint arxiv …, 2024 - arxiv.org
Prepending model inputs with safety prompts is a common practice for safeguarding large
language models (LLMs) against queries with harmful intents. However, the underlying …

Amplegcg: Learning a universal and transferable generative model of adversarial suffixes for jailbreaking both open and closed llms

Z Liao, H Sun - arxiv preprint arxiv:2404.07921, 2024 - arxiv.org
As large language models (LLMs) become increasingly prevalent and integrated into
autonomous systems, ensuring their safety is imperative. Despite significant strides toward …

R-judge: Benchmarking safety risk awareness for llm agents

T Yuan, Z He, L Dong, Y Wang, R Zhao, T **a… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) have exhibited great potential in autonomously completing
tasks across real-world applications. Despite this, these LLM agents introduce unexpected …

Llm-based edge intelligence: A comprehensive survey on architectures, applications, security and trustworthiness

O Friha, MA Ferrag, B Kantarci… - IEEE Open Journal …, 2024 - ieeexplore.ieee.org
The integration of Large Language Models (LLMs) and Edge Intelligence (EI) introduces a
groundbreaking paradigm for intelligent edge devices. With their capacity for human-like …

Grounding and evaluation for large language models: Practical challenges and lessons learned (survey)

K Kenthapadi, M Sameki, A Taly - Proceedings of the 30th ACM SIGKDD …, 2024 - dl.acm.org
With the ongoing rapid adoption of Artificial Intelligence (AI)-based systems in high-stakes
domains, ensuring the trustworthiness, safety, and observability of these systems has …

Evaluating frontier models for dangerous capabilities

M Phuong, M Aitchison, E Catt, S Cogan… - arxiv preprint arxiv …, 2024 - arxiv.org
To understand the risks posed by a new AI system, we must understand what it can and
cannot do. Building on prior work, we introduce a programme of new" dangerous capability" …