Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
The rise and potential of large language model based agents: A survey
For a long time, researchers have sought artificial intelligence (AI) that matches or exceeds
human intelligence. AI agents, which are artificial entities capable of sensing the …
human intelligence. AI agents, which are artificial entities capable of sensing the …
[HTML][HTML] AI deception: A survey of examples, risks, and potential solutions
This paper argues that a range of current AI systems have learned how to deceive humans.
We define deception as the systematic inducement of false beliefs in the pursuit of some …
We define deception as the systematic inducement of false beliefs in the pursuit of some …
Avalon's game of thoughts: Battle against deception through recursive contemplation
Recent breakthroughs in large language models (LLMs) have brought remarkable success
in the field of LLM-as-Agent. Nevertheless, a prevalent assumption is that the information …
in the field of LLM-as-Agent. Nevertheless, a prevalent assumption is that the information …
How Far Are We on the Decision-Making of LLMs? Evaluating LLMs' Gaming Ability in Multi-Agent Environments
Decision-making is a complex process requiring diverse abilities, making it an excellent
framework for evaluating Large Language Models (LLMs). Researchers have examined …
framework for evaluating Large Language Models (LLMs). Researchers have examined …
Llm as a mastermind: A survey of strategic reasoning with large language models
This paper presents a comprehensive survey of the current status and opportunities for
Large Language Models (LLMs) in strategic reasoning, a sophisticated form of reasoning …
Large Language Models (LLMs) in strategic reasoning, a sophisticated form of reasoning …
Evaluating frontier models for dangerous capabilities
To understand the risks posed by a new AI system, we must understand what it can and
cannot do. Building on prior work, we introduce a programme of new" dangerous capability" …
cannot do. Building on prior work, we introduce a programme of new" dangerous capability" …
Put your money where your mouth is: Evaluating strategic planning and execution of llm agents in an auction arena
Recent advancements in Large Language Models (LLMs) showcase advanced reasoning,
yet NLP evaluations often depend on static benchmarks. Evaluating this necessitates …
yet NLP evaluations often depend on static benchmarks. Evaluating this necessitates …
Critical thinking in the age of generative AI
The rapid rise of generative artificial intelligence (GenAI) has prompted a vigorous
discussion about the role this technology should play in the business classroom (Adeshola …
discussion about the role this technology should play in the business classroom (Adeshola …
Large language models can strategically deceive their users when put under pressure
We demonstrate a situation in which Large Language Models, trained to be helpful,
harmless, and honest, can display misaligned behavior and strategically deceive their users …
harmless, and honest, can display misaligned behavior and strategically deceive their users …
Agentlens: Visual analysis for agent behaviors in llm-based autonomous systems
Recently, Large Language Model based Autonomous System (LLMAS) has gained great
popularity for its potential to simulate complicated behaviors of human societies. One of its …
popularity for its potential to simulate complicated behaviors of human societies. One of its …