Täuschung durch künstliche Intelligenz
AI deception: A survey of examples, risks, and potential solutions
https://www.cell.com/patterns/fulltext/S2666-3899(24)00103-X
The bigger picture
AI
systems are already capable of deceiving humans. Deception is the
systematic inducement of false beliefs in others to accomplish some
outcome other than the truth. Large language models and other AI systems
have already learned, from their training, the ability to deceive via
techniques such as manipulation, sycophancy, and cheating the safety
test. AI’s increasing capabilities at deception pose serious risks,
ranging from short-term risks, such as fraud and election tampering, to
long-term risks, such as losing control of AI systems. Proactive
solutions are needed, such as regulatory frameworks to assess AI
deception risks, laws requiring transparency about AI interactions, and
further research into detecting and preventing AI deception. Proactively
addressing the problem of AI deception is crucial to ensure that AI
acts as a beneficial technology that augments rather than destabilizes
human knowledge, discourse, and institutions.
Summary
This
paper argues that a range of current AI systems have learned how to
deceive humans. We define deception as the systematic inducement of
false beliefs in the pursuit of some outcome other than the truth. We
first survey empirical examples of AI deception, discussing both
special-use AI systems (including Meta’s CICERO) and general-purpose AI
systems (including large language models). Next, we detail several risks
from AI deception, such as fraud, election tampering, and losing
control of AI. Finally, we outline several potential solutions: first,
regulatory frameworks should subject AI systems that are capable of
deception to robust risk-assessment requirements; second, policymakers
should implement bot-or-not laws; and finally, policymakers should
prioritize the funding of relevant research, including tools to detect
AI deception and to make AI systems less deceptive. Policymakers,
researchers, and the broader public should work proactively to prevent
AI deception from destabilizing the shared foundations of our society.
05/2024