
Optimising for Trouble – Game Theory and AI Safety | with Jobst Heitzig
What happens when an AI system faithfully follows a flawed goal? In this episode, we explore how even well-designed algorithms can produce dangerous outcomes — from amplifying hate speech to mismanaging infrastructure — simply by optimising a reward function which, like all reward functions, fails to encode all that matters. We discuss the hidden risks of reinforcement learning, why over-optimisation can backfire, and how game theory helps us rethink what it means for AI to act "rationally" in complex, real-world environments.
Jobst Heitzig is a mathematician at the Potsdam Institute for Climate Impact Research and an expert in AI safety and decision design.
Flere episoder fra "Game Changer - the game theory podcast"



Gå ikke glip af nogen episoder af “Game Changer - the game theory podcast” - abonnér på podcasten med gratisapp GetPodcast.








