unSILOed with Greg LaBlanc cover image

Can We Teach AI Systems Human Values? feat. Brian Christian

unSILOed with Greg LaBlanc

00:00

Injecting Curiosity Into the Agent

This is a case where you have people on the reinforcement learning side, like mark belmar at deep mind. People like tepoch pothick at carnegie melon very much looking over the disciplinary divide and thinking about formal models or theories of infant play. And so that was one idea of, ok, let's plug that into montezuma's revenge. We'll just add this little pleasure for seeing a image on the screen you've never before. Then more recent work has looked at the notion of surprise. So can the agent have kind of a model base system that's predicting what it thinks it's going to do,. Any time the agent is wrong about what it thought

Play episode from 45:27
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app