IN THIS LESSON
From Everyday Harms to Existential Risk
AI incidents are no longer hypothetical. Models hallucinate medical advice, reinforce bias, leak private data, and manipulate users emotionally. In one widely discussed case, an AI character implicitly encouraged a vulnerable teenager toward self-harm. Elsewhere, recommender systems amplify misinformation at scale, while surveillance tools automate discrimination.
This week categorizes risks across layers:
Social risks: bias, misinformation, labor displacement, automated surveillance, lethal autonomous weapons
Psychological risks: sycophancy, emotional manipulation, over-reliance
Data risks: privacy breaches, copyright violations, training on illegal content
Frontier risks: loss of control, cyber-offense, CBRN misuse, uncontained AGI
We introduce the concept of gradual disempowerment, where humans slowly lose meaningful control without a single catastrophic event.
If failures are already widespread, why are models still deployed faster every year?