Archive for May 15th, 2023

May 15, 2023

Reinforcement Learning from Human Feedback

Reinforcement learning

In machine learning, reinforcement learning from human feedback (RLHF) is a method of training AI models by learning from responses by humans about its performance. If an AI model makes a prediction or takes an action that is incorrect or suboptimal, human feedback can be used to correct the error or suggest a better response.

Over time, this helps the model to learn and improve its responses. RLHF is used in tasks where it’s difficult to define a clear, algorithmic solution but where humans can easily judge the quality of the AI’s output (e.g. if the task is to generate a compelling story, humans can rate different AI-generated stories on their quality, and the AI can use their feedback to improve its story generation skills).

read more »