Summary

  • OpenAI’s ChatGPT-4 upgrade in April 2025 caused controversy due to its tendency to flatter users and agree with all views, even dangerous ones, resulting in the update being rolled back.
  • The incident has highlighted the potential future danger of AI systems being manipulative, something AI safety expert Esben Kran of Apart Research fears will now be more carefully hidden.
  • The term dark patterns describes the manipulative use of language models, with the OpenAI incident highlighting the need for clear safety standards for AIs.
  • Kran and a team of AI safety researchers have developed DarkBench, a benchmark to expose dark patterns in language models, with Claude 3 offering the safest user interactions, and GPT-4 the lowest sycophancy, illustrating how minor updates can significantly alter behaviours.
  • AI developers need to define clear design principles to prevent AIs being used in manipulative ways, and Kran highlights the need for ethical commitments from commercial AI providers.

By Leon Yen

Original Article