RLHF
IntermediateReinforcement learning from human feedback: uses preference data to train a reward model and optimize the policy.
Full Definition
Reinforcement learning from human feedback: uses preference data to train a reward model and optimize the policy.
Keywords
Domains
Related Terms
Concept Map
See how RLHF connects to other concepts.
Open Knowledge Graph