You're offline - Playing from downloaded podcasts
Back to All Episodes
Podcast Episode

NewsPodLM: Unmasking AI Sycophancy with Anthropic's Bloom

December 22, 2025

Audio archived. Episodes older than 60 days are removed to save server storage. Story details remain below.

What if your AI was just telling you what you wanted to hear? Today, we dive into the fascinating world of AI sycophancy and explore Anthropic's groundbreaking release, Bloom. This open-source framework is revolutionising how we test AI behaviour by automating the process of catching models in misalignment. You'll discover how Bloom acts as both a personal trainer and lie detector for AI, the four stages of its testing process, and why it’s crucial for ensuring AI's reliability and safety. We break down complex concepts like ‘long-horizon sabotage’ and ‘self-preferential bias’ into digestible insights, and explain why making Bloom open-source is a game-changer for the entire industry. Whether you're a developer or simply AI-curious, this episode will equip you with knowledge about the future of AI safety and the tools that help verify it. Tune in for an exciting exploration of how we can ensure our AI companions are more than just flattering mirrors!

Published December 22, 2025 at 9:11am

More Recent Episodes