
Illustrating Reinforcement Learning from Human Feedback (RLHF)
BlueDot Narrated
00:00
RLHF: idea and overview
Perrin Walker defines RLHF and previews its use to align models to human values.
Play episode from 01:32
Transcript

Perrin Walker defines RLHF and previews its use to align models to human values.