RLHF
AI/ML Fundamentals
Reinforcement Learning from Human Feedback
What is RLHF?
Combine supervised learning and RL to fine-tune models using human preferences and reward models for safer/aligned behavior.
Real-World Examples
- •Chatbot alignment to human preferences
Related Terms
Learn more about concepts related to RLHF