RLHF

Intermediate

Reinforcement learning from human feedback: uses preference data to train a reward model and optimize the policy.

Full Definition

Reinforcement learning from human feedback: uses preference data to train a reward model and optimize the policy.

Keywords

Domains

Related Terms

Concept Map

See how RLHF connects to other concepts.

Open Knowledge Graph