Definition
RLHF (Reinforcement Learning from Human Feedback)
Training AI using human preferences to improve response quality.
Full Definition
Reinforcement Learning from Human Feedback (RLHF) is a technique for training AI models using human evaluations of response quality. Human raters compare different model outputs, and these preferences are used to fine-tune the model to generate more preferred responses. RLHF is a key technique behind the helpfulness and alignment of models like ChatGPT. For GEO, understanding RLHF explains why AI systems tend to recommend reputable, well-known brands—these are likely preferred in human evaluations.
Related Terms
Tools & Resources
Monitor Your AI Visibility
See how ChatGPT, Claude, and Perplexity mention your brand.
Free AI Visibility Check