LessWrong (30+ Karma) cover image

“AI #161 Part 1: 80,000 Interviews” by Zvi

LessWrong (30+ Karma)

00:00

LLMs as Unreliable Evaluators

Zvi highlights model evaluation failures, adversarial risks, and Google Gemini integration problems using Christoph's tests.

Play episode from 04:59
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app