Your browser doesn't support HTML5 audio
Definition | : | Reinforcement Learning from Human Feedback |
Category | : | Computing » Artificial Intelligence |
Country/ Region |
: | Worldwide
|
Popularity | : |
|
Type | : |
Initialism
|
Reinforcement Learning from Human Feedback (RLHF) is a type of machine learning method where an agent learns to make decisions by receiving feedback from a human in the form of rewards or punishments.
Note:
An agent is a software or algorithm that is designed to take actions in an environment in order to achieve a specific goal or set of goals.
The full form of RLHF is Reinforcement Learning from Human Feedback
Reinforcement Learning from Human Feedback