The Morality Machine — Episode 3: The Ethics of LLM Poisoning
In this episode of The Morality Machine, we discuss the hidden dangers of LLM poisoning. We uncover how bad actors, propaganda networks, and artists are manipulating AI systems.
In this episode of The Morality Machine, we discuss the hidden dangers of LLM poisoning. We uncover how bad actors, propaganda networks, and artists are manipulating AI systems.
In this episode, we test whether large language models (LLMs) can truly be trusted to provide accurate sources. By comparing responses from GPT-oss-20B and DeepSeek to a politically charged research question, we uncover a pattern of fabricated information in AI models.
The same psychological tricks that persuade humans can also manipulate AI. By appealing to authority, emotion, or even imagined stakes, users have convinced models like ChatGPT to ignore their own safety rules. This raises several ethical questions: when is probing these limits a valuable act of research, and when is it reckless?
We use cookies to improve your experience on our site. By using our site, you consent to cookies.
Manage your cookie preferences below:
Essential cookies enable basic functions and are necessary for the proper function of the website.
These cookies are needed for adding comments on this website.
These cookies are used for managing login functionality on this website.
You can find more information in our Privacy Policy.