RLHF (Reinforcement Learning from Human Feedback) [RLHF]

« Back to Glossary Index

RLHF aligns models like ChatGPT with human preferences using feedback-driven reinforcement learning.

Search the Web

« Back to Glossary Index