RLHF

TERM
RLHF
DEFINITION
Reinforcement Learning from Human Feedback — training AI using human preferences