Reinforcement Learning from Human Feedback (RLHF)
Reading Time: 1 minute
Last updated on November 12, 2024
Reinforcement Learning from Human Feedback is a method for fine tuning LLMs according to human preference data.
Valeriia Kuka
Valeriia Kuka, Head of Content at Learn Prompting, is passionate about making AI and ML accessible. Valeriia previously grew a 60K+ follower AI-focused social media account, earning reposts from Stanford NLP, Amazon Research, Hugging Face, and AI researchers. She has also worked with AI/ML newsletters and global communities with 100K+ members and authored clear and concise explainers and historical articles.