RLHF Tutorial

PAIR Lab

This series of tutorial will explore Reinforcement Learning from Human Feedback (RLHF) – an essential advancement in AI. Starting with alignment (Lecture 1), we will move to Reinforcement Learning (Lectures 2, 3) and then delve into the critical role of Human Feedback (Lecture 4). Further, we will focus on how RLHF works on Large Language Models (LLMs), enhancing their alignment with human values (Lecture 5, 6). Finally, we will discuss other alignment methods in LLMs related to RLHF (Lecture 7, 8).

PreviewDownload
Click Here
Click Here
Click Here
Click Here
Click Here
Click Here
Click Here
Click Here